var/home/core/zuul-output/0000755000175000017500000000000015114027053014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114041460015465 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005035372015114041452017677 0ustar rootrootDec 03 12:37:30 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 12:37:30 crc restorecon[4727]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:30 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 12:37:31 crc restorecon[4727]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 12:37:32 crc kubenswrapper[4990]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 12:37:32 crc kubenswrapper[4990]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 12:37:32 crc kubenswrapper[4990]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 12:37:32 crc kubenswrapper[4990]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 12:37:32 crc kubenswrapper[4990]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 12:37:32 crc kubenswrapper[4990]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.061397 4990 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064686 4990 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064712 4990 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064719 4990 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064724 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064728 4990 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064735 4990 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064739 4990 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064745 4990 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064749 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064755 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064760 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064764 4990 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064768 4990 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064773 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064777 4990 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064781 4990 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064787 4990 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064791 4990 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064796 4990 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064800 4990 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064805 4990 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064809 4990 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064813 4990 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064818 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064822 4990 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064826 4990 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064830 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064834 4990 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064839 4990 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064844 4990 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064849 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064855 4990 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064860 4990 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064866 4990 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064870 4990 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064876 4990 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064882 4990 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064888 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064893 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064897 4990 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064902 4990 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064906 4990 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064910 4990 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064914 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064919 4990 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064923 4990 feature_gate.go:330] unrecognized feature gate: Example Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064928 4990 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064932 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064937 4990 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064941 4990 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064945 4990 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064950 4990 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064957 4990 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064963 4990 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064968 4990 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064974 4990 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064979 4990 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064984 4990 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064988 4990 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.064995 4990 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065000 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065005 4990 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065009 4990 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065013 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065017 4990 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065021 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065026 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065031 4990 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065036 4990 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065042 4990 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.065046 4990 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065342 4990 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065361 4990 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065371 4990 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065383 4990 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065389 4990 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065395 4990 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065403 4990 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065409 4990 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065415 4990 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065420 4990 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065425 4990 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065431 4990 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065436 4990 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065441 4990 flags.go:64] FLAG: --cgroup-root="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065465 4990 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065471 4990 flags.go:64] FLAG: --client-ca-file="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065476 4990 flags.go:64] FLAG: --cloud-config="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065481 4990 flags.go:64] FLAG: --cloud-provider="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065487 4990 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065494 4990 flags.go:64] FLAG: --cluster-domain="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065499 4990 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065505 4990 flags.go:64] FLAG: --config-dir="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065510 4990 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065516 4990 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065523 4990 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065528 4990 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065533 4990 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065538 4990 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065543 4990 flags.go:64] FLAG: --contention-profiling="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065549 4990 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065554 4990 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065560 4990 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065565 4990 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065572 4990 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065577 4990 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065582 4990 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065588 4990 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065593 4990 flags.go:64] FLAG: --enable-server="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065599 4990 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065605 4990 flags.go:64] FLAG: --event-burst="100" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065611 4990 flags.go:64] FLAG: --event-qps="50" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065616 4990 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065621 4990 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065626 4990 flags.go:64] FLAG: --eviction-hard="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065633 4990 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065638 4990 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065645 4990 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065650 4990 flags.go:64] FLAG: --eviction-soft="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065655 4990 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065661 4990 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065666 4990 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065671 4990 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065677 4990 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065682 4990 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065688 4990 flags.go:64] FLAG: --feature-gates="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065694 4990 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065699 4990 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065705 4990 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065710 4990 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065715 4990 flags.go:64] FLAG: --healthz-port="10248" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065721 4990 flags.go:64] FLAG: --help="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065726 4990 flags.go:64] FLAG: --hostname-override="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065731 4990 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065736 4990 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065741 4990 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065747 4990 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065752 4990 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065757 4990 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065762 4990 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065768 4990 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065773 4990 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065778 4990 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065786 4990 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065792 4990 flags.go:64] FLAG: --kube-reserved="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065796 4990 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065801 4990 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065806 4990 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065810 4990 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065816 4990 flags.go:64] FLAG: --lock-file="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065821 4990 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065826 4990 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065831 4990 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065839 4990 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065844 4990 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065849 4990 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065853 4990 flags.go:64] FLAG: --logging-format="text" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065858 4990 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065863 4990 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065868 4990 flags.go:64] FLAG: --manifest-url="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065873 4990 flags.go:64] FLAG: --manifest-url-header="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065880 4990 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065885 4990 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065892 4990 flags.go:64] FLAG: --max-pods="110" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065897 4990 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065903 4990 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065908 4990 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065913 4990 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065919 4990 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065924 4990 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065929 4990 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065942 4990 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065947 4990 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065953 4990 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065959 4990 flags.go:64] FLAG: --pod-cidr="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065963 4990 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065972 4990 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065977 4990 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065983 4990 flags.go:64] FLAG: --pods-per-core="0" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065989 4990 flags.go:64] FLAG: --port="10250" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065994 4990 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.065999 4990 flags.go:64] FLAG: --provider-id="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066005 4990 flags.go:64] FLAG: --qos-reserved="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066011 4990 flags.go:64] FLAG: --read-only-port="10255" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066016 4990 flags.go:64] FLAG: --register-node="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066021 4990 flags.go:64] FLAG: --register-schedulable="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066026 4990 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066036 4990 flags.go:64] FLAG: --registry-burst="10" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066040 4990 flags.go:64] FLAG: --registry-qps="5" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066045 4990 flags.go:64] FLAG: --reserved-cpus="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066050 4990 flags.go:64] FLAG: --reserved-memory="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066056 4990 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066061 4990 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066067 4990 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066072 4990 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066076 4990 flags.go:64] FLAG: --runonce="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066081 4990 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066087 4990 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066092 4990 flags.go:64] FLAG: --seccomp-default="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066097 4990 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066102 4990 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066108 4990 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066113 4990 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066118 4990 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066124 4990 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066129 4990 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066134 4990 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066139 4990 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066144 4990 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066149 4990 flags.go:64] FLAG: --system-cgroups="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066154 4990 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066163 4990 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066168 4990 flags.go:64] FLAG: --tls-cert-file="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066174 4990 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066181 4990 flags.go:64] FLAG: --tls-min-version="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066190 4990 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066195 4990 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066200 4990 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066206 4990 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066211 4990 flags.go:64] FLAG: --v="2" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066218 4990 flags.go:64] FLAG: --version="false" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066225 4990 flags.go:64] FLAG: --vmodule="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066231 4990 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066236 4990 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066365 4990 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066371 4990 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066377 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066382 4990 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066387 4990 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066392 4990 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066396 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066400 4990 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066405 4990 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066409 4990 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066414 4990 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066418 4990 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066422 4990 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066426 4990 feature_gate.go:330] unrecognized feature gate: Example Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066431 4990 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066435 4990 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066440 4990 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066444 4990 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066470 4990 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066475 4990 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066519 4990 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066526 4990 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066533 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066538 4990 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066542 4990 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066546 4990 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066551 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066556 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066561 4990 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066565 4990 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066569 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066573 4990 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066578 4990 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066582 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066586 4990 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066592 4990 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066597 4990 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066602 4990 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066606 4990 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066611 4990 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066616 4990 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066621 4990 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066626 4990 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066630 4990 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066636 4990 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066642 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066647 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066654 4990 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066660 4990 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066666 4990 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066671 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066676 4990 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066688 4990 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066693 4990 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066698 4990 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066702 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066706 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066710 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066715 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066719 4990 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066724 4990 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066728 4990 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066732 4990 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066737 4990 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066742 4990 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066746 4990 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066751 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066755 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066760 4990 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066764 4990 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.066768 4990 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.066782 4990 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.077758 4990 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.078065 4990 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078144 4990 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078152 4990 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078157 4990 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078161 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078165 4990 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078168 4990 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078172 4990 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078175 4990 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078181 4990 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078188 4990 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078192 4990 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078196 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078200 4990 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078205 4990 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078209 4990 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078213 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078217 4990 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078223 4990 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078229 4990 feature_gate.go:330] unrecognized feature gate: Example Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078234 4990 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078237 4990 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078241 4990 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078245 4990 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078248 4990 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078252 4990 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078255 4990 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078259 4990 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078262 4990 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078266 4990 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078270 4990 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078274 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078278 4990 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078283 4990 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078287 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078291 4990 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078295 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078299 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078302 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078306 4990 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078310 4990 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078314 4990 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078320 4990 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078323 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078327 4990 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078331 4990 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078336 4990 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078341 4990 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078345 4990 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078349 4990 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078353 4990 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078357 4990 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078361 4990 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078365 4990 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078369 4990 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078373 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078378 4990 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078383 4990 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078388 4990 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078392 4990 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078396 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078400 4990 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078407 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078411 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078416 4990 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078421 4990 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078426 4990 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078431 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078436 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078440 4990 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078445 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078466 4990 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.078475 4990 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078607 4990 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078616 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078620 4990 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078625 4990 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078630 4990 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078634 4990 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078639 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078643 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078648 4990 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078652 4990 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078656 4990 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078660 4990 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078665 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078669 4990 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078673 4990 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078677 4990 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078681 4990 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078685 4990 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078689 4990 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078692 4990 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078696 4990 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078700 4990 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078706 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078710 4990 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078714 4990 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078718 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078722 4990 feature_gate.go:330] unrecognized feature gate: Example Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078726 4990 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078730 4990 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078733 4990 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078737 4990 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078741 4990 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078744 4990 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078749 4990 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078753 4990 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078757 4990 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078761 4990 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078767 4990 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078771 4990 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078777 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078782 4990 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078786 4990 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078791 4990 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078795 4990 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078801 4990 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078807 4990 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078812 4990 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078817 4990 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078822 4990 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078827 4990 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078832 4990 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078836 4990 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078840 4990 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078845 4990 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078849 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078853 4990 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078857 4990 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078861 4990 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078865 4990 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078869 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078874 4990 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078878 4990 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078884 4990 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078889 4990 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078893 4990 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078898 4990 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078902 4990 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078906 4990 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078910 4990 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078914 4990 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.078919 4990 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.078928 4990 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.079358 4990 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.090520 4990 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.090614 4990 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.091111 4990 server.go:997] "Starting client certificate rotation" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.091136 4990 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.091549 4990 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-05 11:34:00.336934408 +0000 UTC Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.091714 4990 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 790h56m28.245224466s for next certificate rotation Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.098655 4990 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.100474 4990 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.108555 4990 log.go:25] "Validated CRI v1 runtime API" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.161646 4990 log.go:25] "Validated CRI v1 image API" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.163527 4990 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.166779 4990 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-12-32-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.166835 4990 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.183905 4990 manager.go:217] Machine: {Timestamp:2025-12-03 12:37:32.181625031 +0000 UTC m=+0.323536260 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a1063038-6a4c-4d75-baa7-5ae3d76716fc BootID:ca993a98-c7de-432f-983e-69d030f7ca14 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:14:c6:29 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:14:c6:29 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:fa:89:f4 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:54:2b:bd Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:b2:d1:1a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b6:29:af Speed:-1 Mtu:1496} {Name:eth10 MacAddress:16:4e:ee:bf:6f:b1 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:de:82:4b:d6:e9:c9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.184241 4990 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.184696 4990 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.185292 4990 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.185521 4990 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.185562 4990 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.185813 4990 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.185827 4990 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.186055 4990 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.186101 4990 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.186384 4990 state_mem.go:36] "Initialized new in-memory state store" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.186521 4990 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.187334 4990 kubelet.go:418] "Attempting to sync node with API server" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.187363 4990 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.187400 4990 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.187416 4990 kubelet.go:324] "Adding apiserver pod source" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.187429 4990 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.190487 4990 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.190932 4990 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.191112 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.191188 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.191155 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.191258 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.191569 4990 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192108 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192131 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192140 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192148 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192162 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192171 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192181 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192194 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192206 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192216 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192229 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.192238 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.193002 4990 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.193532 4990 server.go:1280] "Started kubelet" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.194408 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.193816 4990 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.193816 4990 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 12:37:32 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.195551 4990 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.197494 4990 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.197518 4990 server.go:460] "Adding debug handlers to kubelet server" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.197569 4990 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:28:30.957868888 +0000 UTC Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.197529 4990 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.198198 4990 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.198223 4990 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.198285 4990 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.198498 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.198866 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.198935 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.198889 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="200ms" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.210485 4990 factory.go:55] Registering systemd factory Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.210528 4990 factory.go:221] Registration of the systemd container factory successfully Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.211069 4990 factory.go:153] Registering CRI-O factory Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.211103 4990 factory.go:221] Registration of the crio container factory successfully Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.211424 4990 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.211487 4990 factory.go:103] Registering Raw factory Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.211511 4990 manager.go:1196] Started watching for new ooms in manager Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.212324 4990 manager.go:319] Starting recovery of all containers Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.211757 4990 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.130:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187db4d421e003d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 12:37:32.193498067 +0000 UTC m=+0.335409306,LastTimestamp:2025-12-03 12:37:32.193498067 +0000 UTC m=+0.335409306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.223929 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224020 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224031 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224043 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224052 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224062 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224072 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224083 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224098 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224107 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224118 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224130 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224139 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224151 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224160 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224174 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224217 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224229 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224240 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224251 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224260 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224269 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224279 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224289 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224300 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224311 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224357 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224369 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224378 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224388 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224425 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224437 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224446 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224471 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224481 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224490 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224501 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224512 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224522 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224533 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224542 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224553 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224564 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224574 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224586 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224596 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224612 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224624 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224635 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224646 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224657 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224667 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224684 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224696 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224708 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224719 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224728 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224738 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224749 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224760 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224769 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224778 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224789 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224798 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224808 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224817 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224828 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224839 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224920 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224936 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224947 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224959 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224971 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224984 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.224997 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225012 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225026 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225036 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225049 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225059 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225071 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225082 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225095 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225105 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225115 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225127 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225136 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225148 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225159 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225171 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225181 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225192 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225203 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225215 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225228 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225238 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225248 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225296 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225309 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225321 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225334 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225345 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225354 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225364 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225379 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225390 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225400 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225409 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225420 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225431 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225459 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225470 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225482 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225493 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225503 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225514 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.225524 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226146 4990 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226175 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226188 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226199 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226211 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226222 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226234 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226244 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226255 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226268 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226277 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226288 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226299 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226309 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226321 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226332 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226342 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226352 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226365 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226377 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226389 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226400 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226411 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226422 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226434 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226464 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226478 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226489 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226499 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226511 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226521 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226531 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226541 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226550 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226559 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226570 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226580 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226589 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226599 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226612 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226622 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226633 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226643 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226656 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226667 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226678 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226688 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226698 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226707 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226717 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226726 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226736 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226746 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226756 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226765 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226775 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226789 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226801 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226810 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226820 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226830 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226840 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226851 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226861 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226872 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226881 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226893 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226914 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226923 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226933 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226944 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226955 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226964 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226974 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226984 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.226996 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227008 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227018 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227030 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227041 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227053 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227063 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227074 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227085 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227097 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227110 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227122 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227139 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227152 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227164 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227176 4990 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227186 4990 reconstruct.go:97] "Volume reconstruction finished" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.227194 4990 reconciler.go:26] "Reconciler: start to sync state" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.236077 4990 manager.go:324] Recovery completed Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.245413 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.247076 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.247127 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.247140 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.247920 4990 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.247937 4990 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.247967 4990 state_mem.go:36] "Initialized new in-memory state store" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.261710 4990 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.263181 4990 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.263216 4990 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.263247 4990 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.263291 4990 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 12:37:32 crc kubenswrapper[4990]: W1203 12:37:32.264509 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.264579 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.299430 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.364441 4990 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.399581 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.399739 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="400ms" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.500632 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.565525 4990 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.600988 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.702146 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.800855 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="800ms" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.802856 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.902957 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.959135 4990 policy_none.go:49] "None policy: Start" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.960730 4990 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 12:37:32 crc kubenswrapper[4990]: I1203 12:37:32.960825 4990 state_mem.go:35] "Initializing new in-memory state store" Dec 03 12:37:32 crc kubenswrapper[4990]: E1203 12:37:32.966514 4990 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.003106 4990 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.057599 4990 manager.go:334] "Starting Device Plugin manager" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.057686 4990 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.057707 4990 server.go:79] "Starting device plugin registration server" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.058297 4990 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.058317 4990 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.058926 4990 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.059028 4990 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.059038 4990 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.064707 4990 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 12:37:33 crc kubenswrapper[4990]: W1203 12:37:33.117020 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.117338 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.159406 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.162433 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.162549 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.162576 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.162626 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.164329 4990 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.130:6443: connect: connection refused" node="crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.195918 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.197922 4990 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 04:33:24.451537745 +0000 UTC Dec 03 12:37:33 crc kubenswrapper[4990]: W1203 12:37:33.198375 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.198509 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:33 crc kubenswrapper[4990]: W1203 12:37:33.206314 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.206396 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.365078 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.366248 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.366331 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.366370 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.366417 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.367133 4990 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.130:6443: connect: connection refused" node="crc" Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.601593 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="1.6s" Dec 03 12:37:33 crc kubenswrapper[4990]: W1203 12:37:33.682909 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.683031 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.766963 4990 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.767168 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.768090 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.768944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769036 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769071 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769480 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769534 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769566 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769587 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769624 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769728 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.769781 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: E1203 12:37:33.769975 4990 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.130:6443: connect: connection refused" node="crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.770728 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.770753 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.770762 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.770848 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771223 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771247 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771260 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771282 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771290 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771980 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.771997 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772007 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772106 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772185 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772193 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772215 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772247 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772274 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772737 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772753 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772761 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.772839 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773128 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773196 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773268 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773334 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773353 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773838 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773868 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.773881 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774032 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774058 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774626 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774682 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774750 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774766 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.774777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847035 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847122 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847178 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847219 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847263 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847344 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847393 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847430 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847477 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847528 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847554 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847583 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847607 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847627 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.847643 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949576 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949690 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949871 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949928 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949949 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949975 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.949980 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950018 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950115 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950089 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950148 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950179 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950089 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950225 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950249 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950259 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950297 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950330 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950348 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950378 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950363 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950409 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950418 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950416 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950441 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950500 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950513 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950547 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950522 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:33 crc kubenswrapper[4990]: I1203 12:37:33.950580 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.108011 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.116367 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.133416 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.150248 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.154592 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:34 crc kubenswrapper[4990]: W1203 12:37:34.177940 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3b2d7bacef8a07496a63121501eac2db7110b5e37021ac42d7552288b9fb28df WatchSource:0}: Error finding container 3b2d7bacef8a07496a63121501eac2db7110b5e37021ac42d7552288b9fb28df: Status 404 returned error can't find the container with id 3b2d7bacef8a07496a63121501eac2db7110b5e37021ac42d7552288b9fb28df Dec 03 12:37:34 crc kubenswrapper[4990]: W1203 12:37:34.191928 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f23ba60023d024f340ba074c822dd7478f7db65c26305b258ac7c429adf74472 WatchSource:0}: Error finding container f23ba60023d024f340ba074c822dd7478f7db65c26305b258ac7c429adf74472: Status 404 returned error can't find the container with id f23ba60023d024f340ba074c822dd7478f7db65c26305b258ac7c429adf74472 Dec 03 12:37:34 crc kubenswrapper[4990]: W1203 12:37:34.193399 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6b8f29ed4be9e47c9292f4f16636a245e63d2fb099fd4cfae3bd4c08ffe6c64a WatchSource:0}: Error finding container 6b8f29ed4be9e47c9292f4f16636a245e63d2fb099fd4cfae3bd4c08ffe6c64a: Status 404 returned error can't find the container with id 6b8f29ed4be9e47c9292f4f16636a245e63d2fb099fd4cfae3bd4c08ffe6c64a Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.194884 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.198736 4990 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 16:57:00.697295207 +0000 UTC Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.198791 4990 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 148h19m26.498506469s for next certificate rotation Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.270046 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"aa42f00713658ab396dac4209a6036c85b3da4fcaf97d712bf0787ce286574a6"} Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.271306 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"033377ea752b1a03fb8c9fc67097b507e26631a235cd03679ba17144050378d2"} Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.272718 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3b2d7bacef8a07496a63121501eac2db7110b5e37021ac42d7552288b9fb28df"} Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.274373 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f23ba60023d024f340ba074c822dd7478f7db65c26305b258ac7c429adf74472"} Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.276100 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6b8f29ed4be9e47c9292f4f16636a245e63d2fb099fd4cfae3bd4c08ffe6c64a"} Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.570297 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.571869 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.571903 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.571912 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:34 crc kubenswrapper[4990]: I1203 12:37:34.571934 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:37:34 crc kubenswrapper[4990]: E1203 12:37:34.572616 4990 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.130:6443: connect: connection refused" node="crc" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.195621 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:35 crc kubenswrapper[4990]: E1203 12:37:35.202355 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="3.2s" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.279287 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1"} Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.280565 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222" exitCode=0 Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.280635 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222"} Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.280654 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.281693 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.281715 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.281724 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.282708 4990 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7cca33b93fb2409ade2ea8a645ef8b1bbcc13c1375eb82b04f4b2440ce5a6205" exitCode=0 Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.282742 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7cca33b93fb2409ade2ea8a645ef8b1bbcc13c1375eb82b04f4b2440ce5a6205"} Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.282870 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.283832 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.283850 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.283858 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.283989 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.287603 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.287626 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.287636 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.289839 4990 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea" exitCode=0 Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.289912 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea"} Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.289950 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.290762 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.290782 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.290790 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.291974 4990 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d" exitCode=0 Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.292001 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d"} Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.292053 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.292879 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.292917 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:35 crc kubenswrapper[4990]: I1203 12:37:35.292929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:35 crc kubenswrapper[4990]: W1203 12:37:35.664016 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:35 crc kubenswrapper[4990]: E1203 12:37:35.664098 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:35 crc kubenswrapper[4990]: W1203 12:37:35.899675 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:35 crc kubenswrapper[4990]: E1203 12:37:35.899754 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.173219 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.174387 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.174432 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.174443 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.174482 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:37:36 crc kubenswrapper[4990]: E1203 12:37:36.174929 4990 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.130:6443: connect: connection refused" node="crc" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.195661 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:36 crc kubenswrapper[4990]: W1203 12:37:36.229086 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:36 crc kubenswrapper[4990]: E1203 12:37:36.229186 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.296011 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18"} Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.296053 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b"} Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.297566 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276"} Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.299066 4990 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bf77b35533e8fa6515c7ef92c6fcfaff0b970e9e77613a0fd35b0ee0674fde1d" exitCode=0 Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.299127 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bf77b35533e8fa6515c7ef92c6fcfaff0b970e9e77613a0fd35b0ee0674fde1d"} Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.324384 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.324396 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0"} Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.325264 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.325311 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.325325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:36 crc kubenswrapper[4990]: I1203 12:37:36.329718 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3"} Dec 03 12:37:36 crc kubenswrapper[4990]: W1203 12:37:36.829073 4990 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:36 crc kubenswrapper[4990]: E1203 12:37:36.829169 4990 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.130:6443: connect: connection refused" logger="UnhandledError" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.195245 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.130:6443: connect: connection refused Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.340394 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d"} Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.340444 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a"} Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.340617 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.341362 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.341394 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.341405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.344122 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa"} Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.344238 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.345053 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.345080 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.345090 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.349655 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5"} Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.349693 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018"} Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.349733 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.349729 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.350692 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.350724 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.350734 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.350885 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.350909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:37 crc kubenswrapper[4990]: I1203 12:37:37.350919 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.355225 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80"} Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.355304 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2"} Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.355274 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.356214 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.356249 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.356262 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.358134 4990 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="83a1a212e442c763dff4d6af6ab15605137d6a0a39136c8152dbb0bafea8b210" exitCode=0 Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.358200 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"83a1a212e442c763dff4d6af6ab15605137d6a0a39136c8152dbb0bafea8b210"} Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.358272 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.358280 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.358377 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.358381 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361124 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361174 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361253 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361315 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361336 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361399 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361478 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.361489 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:38 crc kubenswrapper[4990]: I1203 12:37:38.431192 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.363881 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cd444e84a42b1c5a019351755453f4e0e9f5b36eed682119ce798987b6d67748"} Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.363920 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.363934 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cc5de0265403d4771e8dffdbdfc760d25c6327e46deb7daa4d9bd66341502356"} Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.363964 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.364073 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.364161 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.364946 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.364977 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.364986 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.365669 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.365713 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.365671 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.365757 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.365771 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.365733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.375053 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.375932 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.375961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.375973 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.375998 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:37:39 crc kubenswrapper[4990]: I1203 12:37:39.917352 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:40 crc kubenswrapper[4990]: I1203 12:37:40.371901 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d308224149a7c885216f9787271f9bf5a1332fdf34f2baf4766154f6ff61227"} Dec 03 12:37:40 crc kubenswrapper[4990]: I1203 12:37:40.371980 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:40 crc kubenswrapper[4990]: I1203 12:37:40.371994 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0931e0ac34777645b68e9fdb1f568923e31fcb74cfab2c702299e3513f95b4ac"} Dec 03 12:37:40 crc kubenswrapper[4990]: I1203 12:37:40.372900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:40 crc kubenswrapper[4990]: I1203 12:37:40.372951 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:40 crc kubenswrapper[4990]: I1203 12:37:40.372964 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.231629 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.239060 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.239240 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.240410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.240470 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.240487 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.347915 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.380552 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.380636 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.380552 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.380523 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1a5bedb3d1747152a37ee830202dee50c0c16194742e2bedf57e55886d1435d4"} Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.380777 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.381667 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.381700 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.381711 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.381672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.381735 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.381745 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.382415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.382437 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.382445 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.431074 4990 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 12:37:41 crc kubenswrapper[4990]: I1203 12:37:41.431156 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.382925 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.382944 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.382925 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384142 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384174 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384706 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384750 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:42 crc kubenswrapper[4990]: I1203 12:37:42.384766 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:43 crc kubenswrapper[4990]: E1203 12:37:43.064857 4990 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.277492 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.277657 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.279041 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.279083 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.279096 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.743154 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.743281 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.744415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.744520 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:45 crc kubenswrapper[4990]: I1203 12:37:45.744534 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:46 crc kubenswrapper[4990]: I1203 12:37:46.430825 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:46 crc kubenswrapper[4990]: I1203 12:37:46.431014 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:46 crc kubenswrapper[4990]: I1203 12:37:46.432089 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:46 crc kubenswrapper[4990]: I1203 12:37:46.432129 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:46 crc kubenswrapper[4990]: I1203 12:37:46.432139 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:48 crc kubenswrapper[4990]: I1203 12:37:48.195942 4990 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 12:37:48 crc kubenswrapper[4990]: E1203 12:37:48.380290 4990 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187db4d421e003d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 12:37:32.193498067 +0000 UTC m=+0.335409306,LastTimestamp:2025-12-03 12:37:32.193498067 +0000 UTC m=+0.335409306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 12:37:48 crc kubenswrapper[4990]: E1203 12:37:48.403744 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 03 12:37:48 crc kubenswrapper[4990]: I1203 12:37:48.551866 4990 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 12:37:48 crc kubenswrapper[4990]: I1203 12:37:48.551958 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 12:37:48 crc kubenswrapper[4990]: I1203 12:37:48.574566 4990 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 12:37:48 crc kubenswrapper[4990]: I1203 12:37:48.574637 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.922110 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.922236 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.923026 4990 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.923064 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.923262 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.923292 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.923303 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:49 crc kubenswrapper[4990]: I1203 12:37:49.925937 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.135138 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.135354 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.137436 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.137523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.137548 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.181207 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.297404 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.408734 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.409025 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.409414 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.409444 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.409472 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.409601 4990 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.409655 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.410215 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.410258 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:50 crc kubenswrapper[4990]: I1203 12:37:50.410270 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:51 crc kubenswrapper[4990]: I1203 12:37:51.410942 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:37:51 crc kubenswrapper[4990]: I1203 12:37:51.411914 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:37:51 crc kubenswrapper[4990]: I1203 12:37:51.411960 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:37:51 crc kubenswrapper[4990]: I1203 12:37:51.411970 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:37:51 crc kubenswrapper[4990]: I1203 12:37:51.431798 4990 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 12:37:51 crc kubenswrapper[4990]: I1203 12:37:51.431887 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 12:37:53 crc kubenswrapper[4990]: E1203 12:37:53.065436 4990 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.576304 4990 trace.go:236] Trace[892118661]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 12:37:42.219) (total time: 11356ms): Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[892118661]: ---"Objects listed" error: 11356ms (12:37:53.576) Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[892118661]: [11.356671146s] [11.356671146s] END Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.576626 4990 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.576304 4990 trace.go:236] Trace[1054852931]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 12:37:39.310) (total time: 14266ms): Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[1054852931]: ---"Objects listed" error: 14266ms (12:37:53.576) Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[1054852931]: [14.266054904s] [14.266054904s] END Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.576724 4990 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.577814 4990 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.578084 4990 trace.go:236] Trace[1316634376]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 12:37:40.737) (total time: 12840ms): Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[1316634376]: ---"Objects listed" error: 12840ms (12:37:53.578) Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[1316634376]: [12.840382924s] [12.840382924s] END Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.578100 4990 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.578894 4990 trace.go:236] Trace[2106434203]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 12:37:40.204) (total time: 13374ms): Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[2106434203]: ---"Objects listed" error: 13373ms (12:37:53.578) Dec 03 12:37:53 crc kubenswrapper[4990]: Trace[2106434203]: [13.374059503s] [13.374059503s] END Dec 03 12:37:53 crc kubenswrapper[4990]: I1203 12:37:53.578921 4990 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 12:37:53 crc kubenswrapper[4990]: E1203 12:37:53.579444 4990 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.199264 4990 apiserver.go:52] "Watching apiserver" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202143 4990 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202418 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202758 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202790 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.202842 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202757 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.202842 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202907 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202926 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.202933 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.203177 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.245671 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.247960 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.247964 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.248384 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.248589 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.248703 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.248814 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.249799 4990 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.249838 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.250589 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.252023 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.298890 4990 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.313521 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.326893 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.338218 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.349734 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.362913 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.376436 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382562 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382613 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382641 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382706 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382730 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382751 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382772 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382793 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382816 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382836 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382856 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382878 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.382995 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383022 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383042 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383063 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383086 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383112 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383132 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383153 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383174 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383198 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383218 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383238 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383261 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383280 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383302 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383322 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383344 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383365 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383386 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383407 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383431 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383470 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383442 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383494 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383517 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383538 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383568 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383588 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383610 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383629 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383650 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383669 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383687 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383705 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383725 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383745 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383764 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383784 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383804 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383826 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383847 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383867 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383889 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383912 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383934 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383934 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383955 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.383979 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384000 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384022 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384118 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384387 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384423 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384444 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384482 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384501 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384522 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384543 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384564 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384584 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384599 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384610 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384631 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384653 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384678 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384700 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384721 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384742 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384751 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384785 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384811 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384832 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384853 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384875 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384896 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384917 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384937 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384960 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.384982 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385001 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385022 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385042 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385066 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385088 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385112 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385132 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385155 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385174 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385187 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385195 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385253 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385276 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385297 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385317 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385336 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385354 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385373 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385391 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385408 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385427 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385444 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385468 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385484 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385503 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385519 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385536 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385553 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385573 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385603 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385620 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385638 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385652 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385669 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385685 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385703 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385721 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385737 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385762 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385792 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385817 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385835 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385852 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385870 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385886 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385901 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385919 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385936 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385952 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385968 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.385984 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386001 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386017 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386065 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386085 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386101 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386119 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386136 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386153 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386175 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386191 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386208 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386224 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386215 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386249 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386285 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386311 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386336 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386358 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386378 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386394 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386411 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386428 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386434 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386473 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.386446 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387076 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387092 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387151 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387180 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387208 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387232 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387254 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387278 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387280 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387300 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387332 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387370 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387410 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387436 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387465 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387490 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387521 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387547 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387574 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387602 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387624 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387652 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387677 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387701 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387726 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387750 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387773 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387786 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387796 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387843 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387872 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387877 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387898 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387923 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387947 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.387974 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388001 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388025 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388048 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388072 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388094 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388120 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388143 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388147 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388164 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388184 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388207 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388233 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388291 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388292 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388361 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388396 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388402 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388446 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388478 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388503 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388538 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388571 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388600 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388627 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388653 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388675 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388702 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388726 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388800 4990 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388819 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388834 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388849 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388865 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388879 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388892 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388907 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388922 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388935 4990 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388951 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388965 4990 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388978 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388993 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389006 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389020 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389037 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389051 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389063 4990 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389075 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389088 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389102 4990 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388468 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388553 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390921 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388809 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388951 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.388992 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389126 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.389223 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389410 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389562 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389710 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.389746 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389809 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.389870 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:37:54.889849436 +0000 UTC m=+23.031760775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.389906 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390250 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390613 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390720 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390877 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390892 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.391244 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.391364 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.391545 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.391648 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.391831 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.392084 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.392091 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.392376 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.392531 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.392772 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.392787 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393011 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393034 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393061 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393234 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393290 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393407 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393439 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393742 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393833 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393959 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.393650 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.394111 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.394234 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.406635 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.407094 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.390123 4990 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.409988 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.415966 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.408926 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.407433 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.421645 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.425403 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.425850 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.425899 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.425956 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.426190 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.426286 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.426359 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.426428 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.426435 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.426764 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.427314 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.427507 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.427582 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.427755 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.427943 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428015 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428097 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428056 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428299 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428226 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.428465 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:54.928430378 +0000 UTC m=+23.070341607 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428489 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.428515 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:54.92850625 +0000 UTC m=+23.070417559 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428701 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428796 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.428841 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429061 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429165 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429234 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429408 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429541 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429595 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429605 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429630 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429652 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429653 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429738 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429761 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429794 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429822 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429958 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429957 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.429957 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.430626 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5wptp"] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.430659 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.430943 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.431715 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.431742 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.432243 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.437375 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.442336 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.442425 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443183 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443306 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443382 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443541 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443649 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443790 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.443933 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.444090 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.444172 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.444323 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.444786 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.450508 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.450836 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.450851 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.451339 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.451424 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.452202 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.452561 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.452626 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.453132 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.453581 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.454210 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.455916 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.455957 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.455962 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.456845 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.457129 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.457141 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.457637 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.457670 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.457842 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.458164 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.458226 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.458326 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.458522 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.458626 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.458869 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.458896 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.458908 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.458969 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:54.958949153 +0000 UTC m=+23.100860382 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.462110 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459560 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459055 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459362 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459434 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459659 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459893 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.459893 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.460029 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.462243 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.462254 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.462255 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.462288 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.462341 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:54.96232368 +0000 UTC m=+23.104234909 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.460279 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.460601 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.460795 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.460828 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.461090 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.461135 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.461326 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.461544 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.461842 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.460057 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.462547 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.462828 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.465598 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.466473 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.468996 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.469119 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.469297 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.469952 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.469605 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.470434 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.470783 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.470883 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.470949 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.470995 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.471021 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.471377 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.471973 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.473012 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.473545 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.473624 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.473661 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.474211 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.479120 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.479401 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.480089 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.480543 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.480846 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.474680 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.484909 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.485050 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.485651 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.489889 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.489936 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490043 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490058 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490136 4990 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490149 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490162 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490175 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490186 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490115 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490197 4990 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490302 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490320 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490336 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490350 4990 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490378 4990 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490393 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490409 4990 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490424 4990 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490442 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490485 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490501 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490512 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490523 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490535 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490546 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490559 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490570 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490583 4990 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490594 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490605 4990 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490616 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490630 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490641 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490654 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490665 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490675 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490685 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490696 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490707 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490718 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490730 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490741 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490751 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490761 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490773 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490785 4990 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490796 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490808 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490820 4990 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490880 4990 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490899 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490911 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490925 4990 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490937 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490949 4990 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490961 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490972 4990 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490984 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490996 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491007 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491020 4990 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491032 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491044 4990 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491056 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491067 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491079 4990 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491099 4990 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491114 4990 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491130 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491142 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491156 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491167 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491177 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491187 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491197 4990 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491207 4990 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491219 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491228 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491238 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491247 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491257 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491267 4990 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491280 4990 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491289 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491301 4990 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491312 4990 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491321 4990 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491346 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491362 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491374 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491386 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491397 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491409 4990 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491421 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491432 4990 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491444 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491475 4990 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491486 4990 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491497 4990 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491510 4990 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491523 4990 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491534 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491545 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491557 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491567 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491579 4990 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491590 4990 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491603 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491616 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491627 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.491640 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.512622 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.515168 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-d6l4q"] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.517508 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.515319 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.490180 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.517032 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.518193 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.518212 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.518223 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.518561 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80" exitCode=255 Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.518589 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80"} Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.518233 4990 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526012 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526069 4990 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526129 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526141 4990 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526153 4990 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526173 4990 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526185 4990 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526197 4990 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526209 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526225 4990 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526239 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526253 4990 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526272 4990 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526283 4990 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526294 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526307 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526322 4990 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526334 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526345 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526357 4990 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526392 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526404 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526416 4990 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526431 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526487 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526501 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526513 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526531 4990 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526543 4990 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526567 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526579 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526599 4990 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526613 4990 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526626 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526639 4990 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526655 4990 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526667 4990 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526680 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526694 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526706 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526720 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526746 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526777 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526789 4990 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526801 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526813 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526829 4990 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526842 4990 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526854 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526866 4990 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526882 4990 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526894 4990 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.526905 4990 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.532468 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.534426 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.534426 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.537674 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.540664 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.544820 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.545694 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.554668 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.556436 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.561404 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.564633 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.568017 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.580965 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: W1203 12:37:54.597405 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6443c07892a7ac9c43159f2e2bf392aa2a394a286d381d8eb6bcd31e0591bdf1 WatchSource:0}: Error finding container 6443c07892a7ac9c43159f2e2bf392aa2a394a286d381d8eb6bcd31e0591bdf1: Status 404 returned error can't find the container with id 6443c07892a7ac9c43159f2e2bf392aa2a394a286d381d8eb6bcd31e0591bdf1 Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.619850 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628022 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ebaefa7f-af7e-41d7-b729-cb16c06afa2b-hosts-file\") pod \"node-resolver-5wptp\" (UID: \"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\") " pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628073 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-host\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628103 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kww4\" (UniqueName: \"kubernetes.io/projected/ebaefa7f-af7e-41d7-b729-cb16c06afa2b-kube-api-access-4kww4\") pod \"node-resolver-5wptp\" (UID: \"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\") " pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628124 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkvlm\" (UniqueName: \"kubernetes.io/projected/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-kube-api-access-pkvlm\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628140 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-serviceca\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628185 4990 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628195 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628224 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.628234 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.651948 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.703143 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.705825 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.705847 4990 scope.go:117] "RemoveContainer" containerID="62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729118 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-serviceca\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729170 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ebaefa7f-af7e-41d7-b729-cb16c06afa2b-hosts-file\") pod \"node-resolver-5wptp\" (UID: \"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\") " pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729204 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-host\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729229 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kww4\" (UniqueName: \"kubernetes.io/projected/ebaefa7f-af7e-41d7-b729-cb16c06afa2b-kube-api-access-4kww4\") pod \"node-resolver-5wptp\" (UID: \"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\") " pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729244 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkvlm\" (UniqueName: \"kubernetes.io/projected/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-kube-api-access-pkvlm\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729596 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ebaefa7f-af7e-41d7-b729-cb16c06afa2b-hosts-file\") pod \"node-resolver-5wptp\" (UID: \"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\") " pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.729643 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-host\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.730214 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-serviceca\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.731559 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.741681 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.749217 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kww4\" (UniqueName: \"kubernetes.io/projected/ebaefa7f-af7e-41d7-b729-cb16c06afa2b-kube-api-access-4kww4\") pod \"node-resolver-5wptp\" (UID: \"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\") " pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.751331 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.753103 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkvlm\" (UniqueName: \"kubernetes.io/projected/1144ccf4-9fd2-4705-9e8c-0ddace5364fe-kube-api-access-pkvlm\") pod \"node-ca-d6l4q\" (UID: \"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\") " pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.764827 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.777934 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.789602 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.799588 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.834651 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5wptp" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.835138 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.839286 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d6l4q" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.856654 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: W1203 12:37:54.856807 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebaefa7f_af7e_41d7_b729_cb16c06afa2b.slice/crio-923a1934d7b4f4b3b306cc958721e6ae9b082cac57e32527e8d32ee4ab421bfb WatchSource:0}: Error finding container 923a1934d7b4f4b3b306cc958721e6ae9b082cac57e32527e8d32ee4ab421bfb: Status 404 returned error can't find the container with id 923a1934d7b4f4b3b306cc958721e6ae9b082cac57e32527e8d32ee4ab421bfb Dec 03 12:37:54 crc kubenswrapper[4990]: W1203 12:37:54.875935 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1144ccf4_9fd2_4705_9e8c_0ddace5364fe.slice/crio-1acc26f8282c09a681166ee41b4ebddc37c930c3660c3b2f83f9608647c572d4 WatchSource:0}: Error finding container 1acc26f8282c09a681166ee41b4ebddc37c930c3660c3b2f83f9608647c572d4: Status 404 returned error can't find the container with id 1acc26f8282c09a681166ee41b4ebddc37c930c3660c3b2f83f9608647c572d4 Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.880880 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.894194 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.931168 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.934627 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.934713 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.934752 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.934887 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.934944 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:55.93492874 +0000 UTC m=+24.076839969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.935033 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:37:55.935024892 +0000 UTC m=+24.076936121 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.935081 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: E1203 12:37:54.935112 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:55.935102505 +0000 UTC m=+24.077013734 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.980099 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-85qrd"] Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.980416 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.984538 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.985486 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.985811 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.986059 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 12:37:54 crc kubenswrapper[4990]: I1203 12:37:54.986360 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.002904 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.018915 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.025974 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.034611 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.035963 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7bec8561-741c-4adf-bfce-6f80a0dfb170-mcd-auth-proxy-config\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.036005 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7bec8561-741c-4adf-bfce-6f80a0dfb170-rootfs\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.036026 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7bec8561-741c-4adf-bfce-6f80a0dfb170-proxy-tls\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.036055 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.036094 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z2gt\" (UniqueName: \"kubernetes.io/projected/7bec8561-741c-4adf-bfce-6f80a0dfb170-kube-api-access-9z2gt\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.036117 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036330 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036358 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036371 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036418 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:56.036402792 +0000 UTC m=+24.178314121 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036238 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036482 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036492 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.036517 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:56.036508704 +0000 UTC m=+24.178420033 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.048037 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.057202 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.065467 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.076037 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.086032 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.096742 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.137041 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7bec8561-741c-4adf-bfce-6f80a0dfb170-rootfs\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.137080 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7bec8561-741c-4adf-bfce-6f80a0dfb170-proxy-tls\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.137116 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z2gt\" (UniqueName: \"kubernetes.io/projected/7bec8561-741c-4adf-bfce-6f80a0dfb170-kube-api-access-9z2gt\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.137147 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7bec8561-741c-4adf-bfce-6f80a0dfb170-mcd-auth-proxy-config\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.137168 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7bec8561-741c-4adf-bfce-6f80a0dfb170-rootfs\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.137738 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7bec8561-741c-4adf-bfce-6f80a0dfb170-mcd-auth-proxy-config\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.140662 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7bec8561-741c-4adf-bfce-6f80a0dfb170-proxy-tls\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.158506 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z2gt\" (UniqueName: \"kubernetes.io/projected/7bec8561-741c-4adf-bfce-6f80a0dfb170-kube-api-access-9z2gt\") pod \"machine-config-daemon-85qrd\" (UID: \"7bec8561-741c-4adf-bfce-6f80a0dfb170\") " pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.292805 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:37:55 crc kubenswrapper[4990]: W1203 12:37:55.313105 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bec8561_741c_4adf_bfce_6f80a0dfb170.slice/crio-2797dfe1a5736395ed85b024fcf472a4fc6e17571d5c03680c12c8ac2f55bbcc WatchSource:0}: Error finding container 2797dfe1a5736395ed85b024fcf472a4fc6e17571d5c03680c12c8ac2f55bbcc: Status 404 returned error can't find the container with id 2797dfe1a5736395ed85b024fcf472a4fc6e17571d5c03680c12c8ac2f55bbcc Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.370612 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-9c857"] Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.370994 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.371805 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2kt75"] Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.372416 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.374195 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.374438 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.374639 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.374967 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.375122 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.375231 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.375248 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.380643 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7lrq6"] Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.381563 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.383444 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.383734 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.384608 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.384784 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.385295 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.385464 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.386574 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.389316 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.402317 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.420280 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.430649 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439500 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-os-release\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439551 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dc81041d-8c71-4fcb-8901-8047795b21c2-cni-binary-copy\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439574 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-ovn\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439590 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-netns\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439604 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-etc-kubernetes\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439641 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-system-cni-dir\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439658 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-slash\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439672 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-netns\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439687 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439720 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439752 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-socket-dir-parent\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439792 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-cnibin\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439807 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-systemd\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439855 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-conf-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439873 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439888 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llqxs\" (UniqueName: \"kubernetes.io/projected/dc81041d-8c71-4fcb-8901-8047795b21c2-kube-api-access-llqxs\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439933 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-etc-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439948 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-kubelet\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.439974 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-kubelet\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440011 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-ovn-kubernetes\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440025 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-netd\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440064 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/41a42811-350f-4311-bb49-f7f116aa6952-multus-daemon-config\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440133 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-log-socket\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440185 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-system-cni-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440207 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-k8s-cni-cncf-io\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440253 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-cni-multus\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440275 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2mcv\" (UniqueName: \"kubernetes.io/projected/41a42811-350f-4311-bb49-f7f116aa6952-kube-api-access-w2mcv\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440298 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-cnibin\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440318 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-bin\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440381 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-os-release\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440407 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-var-lib-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440469 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-cni-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440493 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/41a42811-350f-4311-bb49-f7f116aa6952-cni-binary-copy\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440513 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-hostroot\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440537 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-multus-certs\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440571 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dc81041d-8c71-4fcb-8901-8047795b21c2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440599 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-systemd-units\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440617 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-node-log\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.440655 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-cni-bin\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.445527 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.464606 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.471929 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.489236 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.499606 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.513026 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.521979 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.522662 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.524585 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.524866 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.525545 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"2797dfe1a5736395ed85b024fcf472a4fc6e17571d5c03680c12c8ac2f55bbcc"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.526762 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d6l4q" event={"ID":"1144ccf4-9fd2-4705-9e8c-0ddace5364fe","Type":"ContainerStarted","Data":"1acc26f8282c09a681166ee41b4ebddc37c930c3660c3b2f83f9608647c572d4"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.527956 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.527979 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8cc40a2100a440bbcf134e8f4db83d789a158a8fafeaac1e0aebc7dfa792e0ce"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.529225 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5wptp" event={"ID":"ebaefa7f-af7e-41d7-b729-cb16c06afa2b","Type":"ContainerStarted","Data":"4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.529252 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5wptp" event={"ID":"ebaefa7f-af7e-41d7-b729-cb16c06afa2b","Type":"ContainerStarted","Data":"923a1934d7b4f4b3b306cc958721e6ae9b082cac57e32527e8d32ee4ab421bfb"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.530374 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7ebb0fdd5d6517c203a2d5981a78745512206a59cd14136f5b7b45801eb1aafa"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.531468 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.531492 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6443c07892a7ac9c43159f2e2bf392aa2a394a286d381d8eb6bcd31e0591bdf1"} Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.539524 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544173 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42qbm\" (UniqueName: \"kubernetes.io/projected/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-kube-api-access-42qbm\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544236 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-netd\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544265 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/41a42811-350f-4311-bb49-f7f116aa6952-multus-daemon-config\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544331 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-log-socket\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544355 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-system-cni-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544374 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-k8s-cni-cncf-io\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544415 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-cni-multus\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544404 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-netd\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544492 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2mcv\" (UniqueName: \"kubernetes.io/projected/41a42811-350f-4311-bb49-f7f116aa6952-kube-api-access-w2mcv\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544537 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-cnibin\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544582 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-bin\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544618 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-script-lib\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544637 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-var-lib-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544639 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-log-socket\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544653 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-os-release\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544636 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-system-cni-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544673 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-cni-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544676 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-k8s-cni-cncf-io\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544684 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-cni-multus\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544714 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-bin\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544725 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/41a42811-350f-4311-bb49-f7f116aa6952-cni-binary-copy\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544747 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-hostroot\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544770 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovn-node-metrics-cert\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544799 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-cnibin\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544861 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-cni-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544899 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-hostroot\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.544992 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-os-release\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545016 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-var-lib-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545080 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-multus-certs\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545106 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dc81041d-8c71-4fcb-8901-8047795b21c2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545109 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/41a42811-350f-4311-bb49-f7f116aa6952-multus-daemon-config\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545128 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-systemd-units\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545180 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-node-log\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545211 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-systemd-units\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545228 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-multus-certs\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545250 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-node-log\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545279 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-cni-bin\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545305 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-cni-bin\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545324 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dc81041d-8c71-4fcb-8901-8047795b21c2-cni-binary-copy\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545353 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-ovn\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545371 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-config\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545396 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-os-release\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545407 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/41a42811-350f-4311-bb49-f7f116aa6952-cni-binary-copy\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545413 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-netns\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545473 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-os-release\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545486 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-run-netns\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545490 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-etc-kubernetes\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545529 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-etc-kubernetes\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545546 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-system-cni-dir\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545563 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-ovn\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545604 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-slash\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545627 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-netns\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545690 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545695 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-system-cni-dir\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545716 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-netns\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545726 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-slash\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545740 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-env-overrides\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545766 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545795 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-socket-dir-parent\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545808 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dc81041d-8c71-4fcb-8901-8047795b21c2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545822 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545876 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545881 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-socket-dir-parent\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545886 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dc81041d-8c71-4fcb-8901-8047795b21c2-cni-binary-copy\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545936 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-cnibin\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545973 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-systemd\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.545996 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-cnibin\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546072 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-systemd\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546123 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-conf-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546161 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546187 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llqxs\" (UniqueName: \"kubernetes.io/projected/dc81041d-8c71-4fcb-8901-8047795b21c2-kube-api-access-llqxs\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546240 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-kubelet\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546262 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-etc-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546315 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-kubelet\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546337 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-ovn-kubernetes\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546364 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-etc-openvswitch\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546409 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-kubelet\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546439 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-multus-conf-dir\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546605 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-ovn-kubernetes\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.546635 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/41a42811-350f-4311-bb49-f7f116aa6952-host-var-lib-kubelet\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.554582 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.563265 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.563293 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llqxs\" (UniqueName: \"kubernetes.io/projected/dc81041d-8c71-4fcb-8901-8047795b21c2-kube-api-access-llqxs\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.563479 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2mcv\" (UniqueName: \"kubernetes.io/projected/41a42811-350f-4311-bb49-f7f116aa6952-kube-api-access-w2mcv\") pod \"multus-9c857\" (UID: \"41a42811-350f-4311-bb49-f7f116aa6952\") " pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.571050 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.579736 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.592477 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.609463 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.620012 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.623006 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dc81041d-8c71-4fcb-8901-8047795b21c2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2kt75\" (UID: \"dc81041d-8c71-4fcb-8901-8047795b21c2\") " pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.632699 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.644500 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647050 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42qbm\" (UniqueName: \"kubernetes.io/projected/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-kube-api-access-42qbm\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647102 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-script-lib\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647124 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovn-node-metrics-cert\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647146 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-config\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647168 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-env-overrides\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647734 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-env-overrides\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.647953 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-script-lib\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.648065 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-config\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.650199 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovn-node-metrics-cert\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.658650 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.665973 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42qbm\" (UniqueName: \"kubernetes.io/projected/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-kube-api-access-42qbm\") pod \"ovnkube-node-7lrq6\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.670136 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.706283 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.707611 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9c857" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.718992 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2kt75" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.783490 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.783636 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.795362 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.807375 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.815253 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.844263 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.892027 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.926511 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.955372 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.955593 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:37:57.95556534 +0000 UTC m=+26.097476569 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.955646 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.955693 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.955805 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.955858 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:57.955841388 +0000 UTC m=+26.097752617 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.955900 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: E1203 12:37:55.955924 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:57.95591669 +0000 UTC m=+26.097828019 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:55 crc kubenswrapper[4990]: I1203 12:37:55.966029 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.003631 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.048819 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.056607 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.056815 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.056837 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.056849 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.056906 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:58.056888408 +0000 UTC m=+26.198799637 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.056931 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.056952 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.056986 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.057014 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.057091 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:37:58.057067763 +0000 UTC m=+26.198979052 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.086751 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.144598 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.265614 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.265667 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.265734 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.265741 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.265822 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:37:56 crc kubenswrapper[4990]: E1203 12:37:56.265900 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.268325 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.268897 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.270161 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.270976 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.271989 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.272522 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.273098 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.274234 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.275190 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.277928 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.278402 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.279567 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.280085 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.280615 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.281544 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.282058 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.282803 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.283090 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.283556 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.284129 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.287183 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.287730 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.288723 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.289173 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.291676 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.292202 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.292795 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.293914 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.294481 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.295616 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.296118 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.297068 4990 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.297221 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.299245 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.302150 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.302838 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.304416 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.305079 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.306002 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.306641 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.307748 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.308311 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.309337 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.310050 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.311166 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.311723 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.312823 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.313602 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.314964 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.315610 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.316493 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.316947 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.317879 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.318444 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.318931 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.591139 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.592737 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d6l4q" event={"ID":"1144ccf4-9fd2-4705-9e8c-0ddace5364fe","Type":"ContainerStarted","Data":"973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.593966 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerStarted","Data":"6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.593999 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerStarted","Data":"2891ed901ec511d55fbc958e4be8c482684053b8d28521fe0a22981ebbc05c06"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.609839 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.609922 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.614311 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e" exitCode=0 Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.614368 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.614430 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"85ee693ad9ce3c08979c21b865d8709321d746c5acc57d95e081c2fa42b9b3f8"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.622269 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.627315 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerStarted","Data":"eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.627369 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerStarted","Data":"c5682607f016c2e533556aaa76b6058dbdaec1fd6ae050afaabf037acacb6e34"} Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.637393 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.654123 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.664739 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.677178 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.694213 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.712392 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.752622 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.763274 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.775982 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.795113 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.808518 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.828510 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.837728 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.853585 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.867176 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.879640 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.893028 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.931542 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:56 crc kubenswrapper[4990]: I1203 12:37:56.967709 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:56Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.009801 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.047963 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.095272 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.133696 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.201906 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.231950 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.642339 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d"} Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.642678 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258"} Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.642694 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe"} Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.643956 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0"} Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.646014 4990 generic.go:334] "Generic (PLEG): container finished" podID="dc81041d-8c71-4fcb-8901-8047795b21c2" containerID="6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662" exitCode=0 Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.646415 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerDied","Data":"6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662"} Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.662163 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.679020 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.691447 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.707415 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.726668 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.738891 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.749778 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.765570 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.787657 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.806957 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.821246 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.833916 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.847277 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.861611 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.876608 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.891212 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.904685 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.927047 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:57 crc kubenswrapper[4990]: I1203 12:37:57.967744 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:57Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.006705 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.030333 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.030407 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.030466 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.030517 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:38:02.030501958 +0000 UTC m=+30.172413187 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.030583 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.030683 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:02.030660692 +0000 UTC m=+30.172571941 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.030598 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.030745 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:02.030735644 +0000 UTC m=+30.172646873 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.054301 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.101786 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.129591 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.139485 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.139552 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139678 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139692 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139702 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139713 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139760 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139779 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139765 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:02.139752577 +0000 UTC m=+30.281663796 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.139860 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:02.139840909 +0000 UTC m=+30.281752198 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.214749 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.245562 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.266519 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.266640 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.266709 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.266761 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.266810 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:37:58 crc kubenswrapper[4990]: E1203 12:37:58.266862 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.307084 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.443173 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.458479 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.463035 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.468043 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.483928 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.499855 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.527533 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.547418 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.562300 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.598652 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.611367 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.633101 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.656375 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerStarted","Data":"115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278"} Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.660207 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9"} Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.660247 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b"} Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.667085 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.710923 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.792626 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.818679 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.830127 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.874201 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.906951 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.953827 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:58 crc kubenswrapper[4990]: I1203 12:37:58.992054 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:58Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.041812 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.108290 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.122695 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.145017 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.186365 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.229214 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.272397 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.337809 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.368633 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.665057 4990 generic.go:334] "Generic (PLEG): container finished" podID="dc81041d-8c71-4fcb-8901-8047795b21c2" containerID="115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278" exitCode=0 Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.665170 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerDied","Data":"115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278"} Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.670198 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20"} Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.710960 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.729073 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.757000 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.773597 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.789846 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.800303 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.812804 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.827484 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.838595 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.851592 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.864710 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.875775 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.886475 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:37:59 crc kubenswrapper[4990]: I1203 12:37:59.907815 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:37:59Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.006859 4990 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.008913 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.008952 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.008965 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.009078 4990 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.016327 4990 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.016601 4990 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.017484 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.017510 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.017519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.017533 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.017543 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.034665 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.037736 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.037875 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.037950 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.038020 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.038091 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.051328 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.055225 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.055288 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.055301 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.055319 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.055335 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.067487 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.071626 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.071646 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.071654 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.071675 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.071684 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.083888 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.087163 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.087197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.087207 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.087222 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.087231 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.099168 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.099289 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.104197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.104234 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.104244 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.104259 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.104270 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.206390 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.206440 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.206466 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.206484 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.206496 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.263771 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.263780 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.263868 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.263973 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.264046 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:00 crc kubenswrapper[4990]: E1203 12:38:00.264154 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.308884 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.308917 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.308926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.308940 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.308949 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.411323 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.411375 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.411388 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.411406 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.411419 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.513421 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.513474 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.513483 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.513497 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.513506 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.615927 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.615972 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.615984 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.615999 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.616012 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.676068 4990 generic.go:334] "Generic (PLEG): container finished" podID="dc81041d-8c71-4fcb-8901-8047795b21c2" containerID="d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d" exitCode=0 Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.676110 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerDied","Data":"d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.689366 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.704630 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.717882 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.718071 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.718130 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.718188 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.718248 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.719045 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.730528 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.743988 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.759224 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.775556 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.790208 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.802705 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.813745 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.820365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.820402 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.820410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.820423 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.820432 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.825702 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.841039 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.859078 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.870785 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:00Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.922981 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.923020 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.923030 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.923047 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:00 crc kubenswrapper[4990]: I1203 12:38:00.923058 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:00Z","lastTransitionTime":"2025-12-03T12:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.025421 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.025444 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.025468 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.025482 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.025491 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.127064 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.127112 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.127123 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.127140 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.127154 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.230368 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.230620 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.230630 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.230669 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.230678 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.332629 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.332665 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.332672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.332686 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.332700 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.434961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.435004 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.435016 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.435033 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.435044 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.536581 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.536618 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.536629 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.536644 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.536656 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.638337 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.638378 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.638387 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.638399 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.638408 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.681541 4990 generic.go:334] "Generic (PLEG): container finished" podID="dc81041d-8c71-4fcb-8901-8047795b21c2" containerID="6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd" exitCode=0 Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.681580 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerDied","Data":"6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.686809 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.697525 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.711782 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.722576 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.733217 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.739909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.739954 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.739965 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.739983 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.739997 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.749828 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.770057 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.784942 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.796879 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.806537 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.818233 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.829640 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.842725 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.842968 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.842985 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.842993 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.843005 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.843013 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.856052 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.869496 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.945587 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.945627 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.945638 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.945654 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:01 crc kubenswrapper[4990]: I1203 12:38:01.945667 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:01Z","lastTransitionTime":"2025-12-03T12:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.048099 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.048131 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.048139 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.048166 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.048188 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.122675 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.122795 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:38:10.122773057 +0000 UTC m=+38.264684286 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.122853 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.122897 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.123005 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.123051 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:10.123040804 +0000 UTC m=+38.264952043 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.123410 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.123441 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:10.123431834 +0000 UTC m=+38.265343053 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.150691 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.150746 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.150759 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.150774 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.150784 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.224050 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.224117 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224259 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224280 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224293 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224294 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224333 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224350 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224354 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:10.224335561 +0000 UTC m=+38.366246790 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.224426 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:10.224405713 +0000 UTC m=+38.366317022 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.254235 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.254284 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.254295 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.254311 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.254323 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.263541 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.263552 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.263680 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.263552 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.263761 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:02 crc kubenswrapper[4990]: E1203 12:38:02.263828 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.275485 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.292926 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.307413 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.322371 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.335212 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.348799 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.357041 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.357107 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.357124 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.357148 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.357168 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.363085 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.379846 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.394633 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.406128 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.415181 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.428390 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.439829 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.450719 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.459305 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.459357 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.459373 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.459390 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.459401 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.561390 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.561432 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.561444 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.561484 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.561496 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.663999 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.664041 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.664051 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.664065 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.664074 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.766677 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.766710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.766720 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.766733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.766744 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.869342 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.869382 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.869391 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.869406 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.869415 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.971609 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.971660 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.971688 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.971714 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:02 crc kubenswrapper[4990]: I1203 12:38:02.971730 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:02Z","lastTransitionTime":"2025-12-03T12:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.072756 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.072793 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.072805 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.072819 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.072831 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.174533 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.174583 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.174597 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.174615 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.174630 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.277423 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.277491 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.277503 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.277519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.277530 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.380293 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.380578 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.380705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.380803 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.380892 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.483533 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.483811 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.483901 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.483996 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.484099 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.587320 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.587369 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.587380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.587396 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.587405 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.689898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.689952 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.689967 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.689985 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.689996 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.792264 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.792332 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.792347 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.792366 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.792382 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.895210 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.895272 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.895288 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.895311 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.895329 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.998120 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.998161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.998170 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.998187 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:03 crc kubenswrapper[4990]: I1203 12:38:03.998198 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:03Z","lastTransitionTime":"2025-12-03T12:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.101489 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.101916 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.101926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.101942 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.101954 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.207941 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.207978 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.207989 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.208004 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.208015 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.265663 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:04 crc kubenswrapper[4990]: E1203 12:38:04.265767 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.266158 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:04 crc kubenswrapper[4990]: E1203 12:38:04.266216 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.266257 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:04 crc kubenswrapper[4990]: E1203 12:38:04.266310 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.309785 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.309814 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.309823 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.309834 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.309842 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.440887 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.440932 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.440945 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.440961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.440972 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.543321 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.543365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.543377 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.543395 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.543407 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.646704 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.646743 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.646756 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.646770 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.646781 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.703101 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.703699 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.703720 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.712889 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerStarted","Data":"6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.722318 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.748412 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.748473 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.748490 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.748510 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.748521 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.752395 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.755386 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.755995 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.765350 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.778066 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.792805 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.813196 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.830157 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.845243 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.850631 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.850676 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.850687 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.850705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.850716 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.857284 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.869423 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.881783 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.892513 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.903432 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.919122 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.933165 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.946124 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.958671 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.958715 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.958725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.958768 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.958786 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:04Z","lastTransitionTime":"2025-12-03T12:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.960076 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.981966 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:04 crc kubenswrapper[4990]: I1203 12:38:04.995802 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:04Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.008314 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.020233 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.031051 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.047580 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.060725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.060772 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.060783 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.060797 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.060806 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.068820 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.081037 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.092044 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.104110 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.118129 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:05Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.162856 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.162892 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.162900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.162916 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.162926 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.264997 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.265058 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.265073 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.265095 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.265109 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.367877 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.367918 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.367929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.367945 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.367954 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.470426 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.470494 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.470507 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.470523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.470534 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.573000 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.573062 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.573086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.573107 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.573120 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.675924 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.675979 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.675993 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.676010 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.676020 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.715058 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.777863 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.777909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.777922 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.777940 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.777950 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.880992 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.881024 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.881032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.881044 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.881054 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.983842 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.983894 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.983907 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.983926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:05 crc kubenswrapper[4990]: I1203 12:38:05.983938 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:05Z","lastTransitionTime":"2025-12-03T12:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.086960 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.087003 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.087015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.087032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.087043 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.189748 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.189787 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.189795 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.189812 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.189826 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.264355 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.264355 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.264374 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:06 crc kubenswrapper[4990]: E1203 12:38:06.264559 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:06 crc kubenswrapper[4990]: E1203 12:38:06.264698 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:06 crc kubenswrapper[4990]: E1203 12:38:06.264799 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.292724 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.292766 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.292774 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.292789 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.292798 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.395055 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.395095 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.395104 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.395120 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.395130 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.497233 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.497277 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.497288 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.497304 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.497317 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.599986 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.600022 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.600030 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.600044 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.600054 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.702814 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.702860 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.702872 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.702992 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.703014 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.721263 4990 generic.go:334] "Generic (PLEG): container finished" podID="dc81041d-8c71-4fcb-8901-8047795b21c2" containerID="6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9" exitCode=0 Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.721345 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerDied","Data":"6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.721396 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.735765 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.751535 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.767064 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.781900 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.796330 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.805415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.805463 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.805471 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.805486 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.805494 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.809753 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.825202 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.849145 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.862971 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.876474 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.888529 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.899903 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.908550 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.908589 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.908602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.908619 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.908630 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:06Z","lastTransitionTime":"2025-12-03T12:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.914746 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:06 crc kubenswrapper[4990]: I1203 12:38:06.926218 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:06Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.010593 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.010636 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.010645 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.010661 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.010670 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.113260 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.113300 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.113309 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.113323 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.113332 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.215063 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.215109 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.215122 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.215139 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.215152 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.317707 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.317744 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.317755 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.317769 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.317780 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.420982 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.421032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.421042 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.421084 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.421096 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.523275 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.523315 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.523325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.523340 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.523349 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.625830 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.625896 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.625910 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.625937 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.625955 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.728009 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.728123 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.728148 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.728177 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.728203 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.831630 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.831689 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.831702 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.831721 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.831732 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.934375 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.934418 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.934426 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.934441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.934475 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:07Z","lastTransitionTime":"2025-12-03T12:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.986572 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk"] Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.987109 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.989622 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 12:38:07 crc kubenswrapper[4990]: I1203 12:38:07.990149 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.003304 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.019876 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.030271 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.036944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.036994 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.037003 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.037017 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.037027 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.041331 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.055176 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.074205 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.088754 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.101388 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.111656 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtdfs\" (UniqueName: \"kubernetes.io/projected/ba784e45-d065-42c9-8b4d-f22c277d8e6c-kube-api-access-jtdfs\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.111846 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba784e45-d065-42c9-8b4d-f22c277d8e6c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.111881 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba784e45-d065-42c9-8b4d-f22c277d8e6c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.111906 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba784e45-d065-42c9-8b4d-f22c277d8e6c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.112289 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.124121 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.135515 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.138816 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.138856 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.138867 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.138883 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.138892 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.149616 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.164209 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.176314 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.192346 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:08Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.212587 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtdfs\" (UniqueName: \"kubernetes.io/projected/ba784e45-d065-42c9-8b4d-f22c277d8e6c-kube-api-access-jtdfs\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.212661 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba784e45-d065-42c9-8b4d-f22c277d8e6c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.212695 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba784e45-d065-42c9-8b4d-f22c277d8e6c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.212714 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba784e45-d065-42c9-8b4d-f22c277d8e6c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.213605 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba784e45-d065-42c9-8b4d-f22c277d8e6c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.213743 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba784e45-d065-42c9-8b4d-f22c277d8e6c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.220054 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba784e45-d065-42c9-8b4d-f22c277d8e6c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.230805 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtdfs\" (UniqueName: \"kubernetes.io/projected/ba784e45-d065-42c9-8b4d-f22c277d8e6c-kube-api-access-jtdfs\") pod \"ovnkube-control-plane-749d76644c-bjsxk\" (UID: \"ba784e45-d065-42c9-8b4d-f22c277d8e6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.241274 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.241329 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.241339 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.241358 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.241369 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.264665 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.264786 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.264665 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:08 crc kubenswrapper[4990]: E1203 12:38:08.264927 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:08 crc kubenswrapper[4990]: E1203 12:38:08.265065 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:08 crc kubenswrapper[4990]: E1203 12:38:08.265163 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.300674 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.343566 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.343607 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.343617 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.343631 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.343642 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.446688 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.446716 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.446726 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.446742 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.446754 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.550093 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.550150 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.550163 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.550180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.550191 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.653149 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.653184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.653193 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.653207 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.653216 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.730795 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerStarted","Data":"84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.731695 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" event={"ID":"ba784e45-d065-42c9-8b4d-f22c277d8e6c","Type":"ContainerStarted","Data":"9261219ea70f9ede63ce292de212884626aef18058f38ee590bb3cac29dfbcb4"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.755743 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.755774 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.755782 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.755795 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.755805 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.858325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.858362 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.858374 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.858391 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.858404 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.962313 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.962361 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.962378 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.962398 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:08 crc kubenswrapper[4990]: I1203 12:38:08.962410 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:08Z","lastTransitionTime":"2025-12-03T12:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.064320 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.064365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.064379 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.064405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.064418 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.089156 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-6sqk8"] Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.089702 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: E1203 12:38:09.089795 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.105407 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.118729 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.134192 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.146565 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.157334 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.166841 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.166871 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.166879 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.166892 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.166902 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.171252 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.190248 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.201072 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.212103 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.221575 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.221741 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb9nh\" (UniqueName: \"kubernetes.io/projected/a061fb3f-ace6-4ab6-a499-b34979186d08-kube-api-access-zb9nh\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.225392 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.235523 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.249270 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.259962 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.268980 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.269017 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.269027 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.269039 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.269049 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.270882 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.283480 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.296722 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.323367 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb9nh\" (UniqueName: \"kubernetes.io/projected/a061fb3f-ace6-4ab6-a499-b34979186d08-kube-api-access-zb9nh\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.323427 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: E1203 12:38:09.323607 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:09 crc kubenswrapper[4990]: E1203 12:38:09.323675 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:09.823657464 +0000 UTC m=+37.965568693 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.340837 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb9nh\" (UniqueName: \"kubernetes.io/projected/a061fb3f-ace6-4ab6-a499-b34979186d08-kube-api-access-zb9nh\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.371346 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.371408 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.371420 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.371438 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.371447 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.474287 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.474333 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.474344 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.474358 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.474368 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.576741 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.576800 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.576817 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.576842 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.576860 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.679177 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.679204 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.679212 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.679224 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.679232 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.737639 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" event={"ID":"ba784e45-d065-42c9-8b4d-f22c277d8e6c","Type":"ContainerStarted","Data":"be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.740942 4990 generic.go:334] "Generic (PLEG): container finished" podID="dc81041d-8c71-4fcb-8901-8047795b21c2" containerID="84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652" exitCode=0 Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.741018 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerDied","Data":"84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.754022 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.767556 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.782031 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.782312 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.782354 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.782365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.782384 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.782399 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.798877 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.815661 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.827679 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:09 crc kubenswrapper[4990]: E1203 12:38:09.828499 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:09 crc kubenswrapper[4990]: E1203 12:38:09.828571 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:10.828551488 +0000 UTC m=+38.970462727 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.829560 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.843019 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.854791 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.866953 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.877599 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.884918 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.884952 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.884965 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.884983 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.884995 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.890561 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.904719 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.924793 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.939565 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.953246 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.965215 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:09Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.987084 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.987119 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.987129 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.987142 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:09 crc kubenswrapper[4990]: I1203 12:38:09.987151 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:09Z","lastTransitionTime":"2025-12-03T12:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.089568 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.089894 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.089907 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.089922 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.089934 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.130342 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.130562 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:38:26.130526968 +0000 UTC m=+54.272438207 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.130680 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.130724 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.130835 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.130855 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.130915 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:26.130894278 +0000 UTC m=+54.272805587 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.130941 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:26.130931099 +0000 UTC m=+54.272842328 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.193067 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.193101 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.193111 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.193125 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.193136 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.231918 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.232014 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232091 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232129 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232131 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232145 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232152 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232156 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232203 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:26.232189954 +0000 UTC m=+54.374101183 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.232219 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:26.232211795 +0000 UTC m=+54.374123034 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.263504 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.263534 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.263570 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.263662 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.263697 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.263803 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.263864 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.263979 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.296125 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.296166 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.296176 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.296194 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.296204 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.389368 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.389421 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.389434 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.389485 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.389502 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.406546 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:10Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.411284 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.411555 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.411643 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.411731 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.411821 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.426649 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:10Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.430899 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.430938 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.430946 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.430963 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.430973 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.445054 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:10Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.449713 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.449749 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.449761 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.449777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.449789 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.463608 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:10Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.467764 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.467802 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.467811 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.467826 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.467836 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.480040 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:10Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.480165 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.481909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.481927 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.481936 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.481950 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.481960 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.584443 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.584497 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.584509 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.584525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.584534 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.686610 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.686650 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.686660 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.686676 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.686687 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.746058 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" event={"ID":"ba784e45-d065-42c9-8b4d-f22c277d8e6c","Type":"ContainerStarted","Data":"df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.789414 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.789468 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.789482 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.789496 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.789506 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.838082 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.838211 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: E1203 12:38:10.838338 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:12.838325702 +0000 UTC m=+40.980236931 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.891782 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.891821 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.891835 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.891855 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.891867 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.993953 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.994001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.994010 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.994024 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:10 crc kubenswrapper[4990]: I1203 12:38:10.994032 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:10Z","lastTransitionTime":"2025-12-03T12:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.096115 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.096167 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.096182 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.096200 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.096213 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.198523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.198563 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.198573 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.198587 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.198597 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.301484 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.301523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.301532 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.301547 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.301557 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.405664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.405720 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.405735 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.405758 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.405788 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.510031 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.510122 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.510138 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.510163 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.510179 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.613035 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.613105 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.613117 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.613157 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.613171 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.714860 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.714902 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.714914 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.714931 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.714946 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.751960 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/0.log" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.754947 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217" exitCode=1 Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.755037 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.755992 4990 scope.go:117] "RemoveContainer" containerID="bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.761635 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" event={"ID":"dc81041d-8c71-4fcb-8901-8047795b21c2","Type":"ContainerStarted","Data":"20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.774409 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.793859 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.811389 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.817534 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.817586 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.817597 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.817616 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.817631 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.831366 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.848045 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.865389 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.878832 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.893411 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.909681 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.920424 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.920474 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.920492 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.920509 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.920519 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:11Z","lastTransitionTime":"2025-12-03T12:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.925831 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.940650 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.952857 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.967284 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:11 crc kubenswrapper[4990]: I1203 12:38:11.987821 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:11Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.007547 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"message\\\":\\\"tSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186395 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186613 6208 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 12:38:11.186661 6208 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 12:38:11.186912 6208 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 12:38:11.186971 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 12:38:11.186978 6208 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 12:38:11.186993 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 12:38:11.187026 6208 factory.go:656] Stopping watch factory\\\\nI1203 12:38:11.187042 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1203 12:38:11.187069 6208 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 12:38:11.187078 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 12:38:11.187083 6208 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 12:38:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.019409 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.022949 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.022991 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.023001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.023017 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.023027 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.031401 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.045083 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.058509 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.079840 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.107998 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.124884 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.125841 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.125880 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.125894 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.125912 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.125924 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.140823 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.155960 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.175239 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.194561 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"message\\\":\\\"tSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186395 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186613 6208 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 12:38:11.186661 6208 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 12:38:11.186912 6208 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 12:38:11.186971 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 12:38:11.186978 6208 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 12:38:11.186993 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 12:38:11.187026 6208 factory.go:656] Stopping watch factory\\\\nI1203 12:38:11.187042 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1203 12:38:11.187069 6208 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 12:38:11.187078 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 12:38:11.187083 6208 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 12:38:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.207082 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.221174 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.228404 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.228468 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.228480 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.228497 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.228509 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.234517 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.246624 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.258692 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.263847 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.263909 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.263906 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.263864 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:12 crc kubenswrapper[4990]: E1203 12:38:12.263999 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:12 crc kubenswrapper[4990]: E1203 12:38:12.264104 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:12 crc kubenswrapper[4990]: E1203 12:38:12.264174 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:12 crc kubenswrapper[4990]: E1203 12:38:12.264220 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.275231 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.288506 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.302840 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.313429 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.329337 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.330671 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.330711 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.330720 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.330733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.330743 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.342990 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.356933 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.369560 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.382697 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.398222 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.419087 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.432917 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.432962 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.432971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.432987 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.432997 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.435888 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.447527 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.460157 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.476426 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.497891 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"message\\\":\\\"tSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186395 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186613 6208 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 12:38:11.186661 6208 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 12:38:11.186912 6208 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 12:38:11.186971 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 12:38:11.186978 6208 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 12:38:11.186993 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 12:38:11.187026 6208 factory.go:656] Stopping watch factory\\\\nI1203 12:38:11.187042 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1203 12:38:11.187069 6208 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 12:38:11.187078 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 12:38:11.187083 6208 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 12:38:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.510633 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.535256 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.535300 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.535310 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.535325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.535337 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.638301 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.638336 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.638346 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.638358 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.638370 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.741075 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.741125 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.741137 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.741153 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.741163 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.768402 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/0.log" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.773239 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.773810 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.790202 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.804865 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.824041 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"message\\\":\\\"tSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186395 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186613 6208 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 12:38:11.186661 6208 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 12:38:11.186912 6208 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 12:38:11.186971 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 12:38:11.186978 6208 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 12:38:11.186993 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 12:38:11.187026 6208 factory.go:656] Stopping watch factory\\\\nI1203 12:38:11.187042 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1203 12:38:11.187069 6208 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 12:38:11.187078 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 12:38:11.187083 6208 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 12:38:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.837678 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.843431 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.843491 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.843502 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.843518 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.843530 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.852264 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.860719 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:12 crc kubenswrapper[4990]: E1203 12:38:12.861030 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:12 crc kubenswrapper[4990]: E1203 12:38:12.861293 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:16.861265895 +0000 UTC m=+45.003177124 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.867632 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.880677 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.894696 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.906210 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.917683 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.931891 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.943527 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.945583 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.945617 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.945628 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.945643 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.945653 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:12Z","lastTransitionTime":"2025-12-03T12:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.958811 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.972783 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:12 crc kubenswrapper[4990]: I1203 12:38:12.985764 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.000028 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:12Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.048577 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.049211 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.049313 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.049424 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.049546 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.151785 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.151820 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.151832 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.151850 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.151862 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.254296 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.254341 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.254351 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.254365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.254376 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.356313 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.356587 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.356658 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.356736 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.356805 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.459797 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.459848 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.459876 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.459894 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.459907 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.547814 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.562432 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.562492 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.562508 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.562526 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.562540 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.671393 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.671504 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.671520 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.671547 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.671562 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.774285 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.774325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.774335 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.774371 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.774384 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.779140 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/1.log" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.779736 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/0.log" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.782421 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1" exitCode=1 Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.782486 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.782581 4990 scope.go:117] "RemoveContainer" containerID="bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.783255 4990 scope.go:117] "RemoveContainer" containerID="3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1" Dec 03 12:38:13 crc kubenswrapper[4990]: E1203 12:38:13.783438 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.796771 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.810902 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.822640 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.838801 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.851632 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.864052 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.876556 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.876599 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.876610 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.876626 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.876639 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.878559 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.889170 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.902509 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.913694 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.924266 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.937316 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.953261 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.972022 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"message\\\":\\\"tSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186395 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186613 6208 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 12:38:11.186661 6208 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 12:38:11.186912 6208 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 12:38:11.186971 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 12:38:11.186978 6208 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 12:38:11.186993 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 12:38:11.187026 6208 factory.go:656] Stopping watch factory\\\\nI1203 12:38:11.187042 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1203 12:38:11.187069 6208 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 12:38:11.187078 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 12:38:11.187083 6208 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 12:38:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.980142 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.980186 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.980198 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.980213 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.980225 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:13Z","lastTransitionTime":"2025-12-03T12:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:13 crc kubenswrapper[4990]: I1203 12:38:13.986589 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.000770 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:13Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.083529 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.083578 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.083606 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.083623 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.083645 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.186608 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.186662 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.186672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.186692 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.186704 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.253206 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.263932 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.263952 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.263969 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.263966 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:14 crc kubenswrapper[4990]: E1203 12:38:14.264977 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:14 crc kubenswrapper[4990]: E1203 12:38:14.265089 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.265108 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: E1203 12:38:14.265152 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:14 crc kubenswrapper[4990]: E1203 12:38:14.265282 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.281156 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.289096 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.289137 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.289151 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.289170 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.289182 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.295957 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.314518 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcccd52ad46b3544fe3c7b1c807699fab66126eca18afde3e9b85c8c10f0b217\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"message\\\":\\\"tSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186395 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 12:38:11.186613 6208 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 12:38:11.186661 6208 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 12:38:11.186912 6208 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 12:38:11.186971 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 12:38:11.186978 6208 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 12:38:11.186993 6208 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 12:38:11.187026 6208 factory.go:656] Stopping watch factory\\\\nI1203 12:38:11.187042 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1203 12:38:11.187069 6208 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 12:38:11.187078 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 12:38:11.187083 6208 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 12:38:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.325490 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.338238 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.352192 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.370340 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.385264 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.394506 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.394559 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.394569 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.394585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.394595 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.401628 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.413895 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.426951 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.441077 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.456092 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.468353 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.481644 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.497378 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.497414 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.497425 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.497441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.497472 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.599421 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.599511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.599524 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.599541 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.599553 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.702444 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.702495 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.702505 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.702519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.702530 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.787409 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/1.log" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.790253 4990 scope.go:117] "RemoveContainer" containerID="3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1" Dec 03 12:38:14 crc kubenswrapper[4990]: E1203 12:38:14.790395 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.800822 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.804079 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.804101 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.804109 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.804124 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.804133 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.812915 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.823534 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.837987 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.849568 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.860428 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.873652 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.889547 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.906499 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.906540 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.906551 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.906565 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.906575 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:14Z","lastTransitionTime":"2025-12-03T12:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.908634 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.918471 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.931762 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.942430 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.951378 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.960996 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.972339 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:14 crc kubenswrapper[4990]: I1203 12:38:14.983856 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:14Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.008419 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.008489 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.008500 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.008516 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.008527 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.110510 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.110575 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.110588 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.110603 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.110613 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.213118 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.213183 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.213200 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.213224 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.213241 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.315550 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.315600 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.315615 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.315637 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.315651 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.417799 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.417860 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.417877 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.417896 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.417909 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.520674 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.520721 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.520736 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.520755 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.520768 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.623614 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.623683 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.623699 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.623721 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.623737 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.727663 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.727723 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.727741 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.727763 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.727784 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.830435 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.830500 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.830513 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.830529 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.830541 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.933275 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.933343 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.933366 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.933394 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:15 crc kubenswrapper[4990]: I1203 12:38:15.933417 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:15Z","lastTransitionTime":"2025-12-03T12:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.035839 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.035898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.035921 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.035949 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.035971 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.138350 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.138377 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.138388 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.138402 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.138413 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.240652 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.240683 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.240698 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.240718 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.240734 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.264447 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.264469 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.264538 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.264769 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:16 crc kubenswrapper[4990]: E1203 12:38:16.264756 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:16 crc kubenswrapper[4990]: E1203 12:38:16.264841 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:16 crc kubenswrapper[4990]: E1203 12:38:16.265049 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:16 crc kubenswrapper[4990]: E1203 12:38:16.265186 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.342971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.343015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.343026 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.343044 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.343057 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.444874 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.444930 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.444941 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.444955 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.444964 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.547757 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.547810 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.547824 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.547840 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.547851 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.650316 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.650368 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.650379 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.650393 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.650401 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.753241 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.753284 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.753293 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.753309 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.753321 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.855070 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.855106 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.855116 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.855130 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.855139 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.900124 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:16 crc kubenswrapper[4990]: E1203 12:38:16.900295 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:16 crc kubenswrapper[4990]: E1203 12:38:16.900351 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:24.900333241 +0000 UTC m=+53.042244460 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.957551 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.957597 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.957606 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.957621 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:16 crc kubenswrapper[4990]: I1203 12:38:16.957632 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:16Z","lastTransitionTime":"2025-12-03T12:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.059871 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.059924 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.059938 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.059957 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.059971 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.161792 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.161846 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.161859 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.161876 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.161888 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.264786 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.264826 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.264836 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.264857 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.264868 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.367587 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.367637 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.367649 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.367663 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.367676 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.469987 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.470046 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.470057 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.470075 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.470088 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.572485 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.572532 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.572544 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.572560 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.572574 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.675601 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.675671 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.675693 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.675716 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.675730 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.778584 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.778633 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.778644 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.778662 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.778674 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.881624 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.881656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.881665 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.881681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.881691 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.984407 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.984470 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.984480 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.984496 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:17 crc kubenswrapper[4990]: I1203 12:38:17.984513 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:17Z","lastTransitionTime":"2025-12-03T12:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.088264 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.089124 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.089164 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.089185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.089196 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.191861 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.191901 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.191911 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.191929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.191941 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.263780 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.263983 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:18 crc kubenswrapper[4990]: E1203 12:38:18.264024 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.264113 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.264208 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:18 crc kubenswrapper[4990]: E1203 12:38:18.264359 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:18 crc kubenswrapper[4990]: E1203 12:38:18.264503 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:18 crc kubenswrapper[4990]: E1203 12:38:18.264589 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.295196 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.295243 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.295254 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.295270 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.295280 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.397762 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.397826 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.397836 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.397850 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.397859 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.500033 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.500074 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.500085 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.500100 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.500111 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.603588 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.603874 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.603971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.604064 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.604152 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.706670 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.706719 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.706735 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.706758 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.706775 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.808371 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.808441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.808475 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.808493 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.808503 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.910574 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.910616 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.910624 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.910640 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:18 crc kubenswrapper[4990]: I1203 12:38:18.910649 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:18Z","lastTransitionTime":"2025-12-03T12:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.012428 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.012507 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.012527 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.012555 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.012567 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.115085 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.115139 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.115153 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.115170 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.115183 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.217986 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.218036 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.218048 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.218068 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.218084 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.321597 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.321651 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.321666 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.321686 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.321697 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.424486 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.424523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.424535 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.424564 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.424577 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.526827 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.526891 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.526905 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.526929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.526942 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.629353 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.629683 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.629767 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.629853 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.629923 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.733011 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.733070 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.733079 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.733102 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.733113 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.836134 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.836190 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.836199 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.836221 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.836233 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.939744 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.939791 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.939802 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.939827 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:19 crc kubenswrapper[4990]: I1203 12:38:19.939840 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:19Z","lastTransitionTime":"2025-12-03T12:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.042431 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.042504 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.042517 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.042533 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.042545 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.146614 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.146679 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.146689 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.146714 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.146728 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.249492 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.249555 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.249569 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.249589 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.249603 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.264371 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.264411 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.264411 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.264444 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.265602 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.265739 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.265856 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.266132 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.357162 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.357262 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.357274 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.357298 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.357310 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.459758 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.459806 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.459818 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.459833 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.459844 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.563484 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.563523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.563532 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.563547 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.563556 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.666204 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.666239 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.666248 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.666262 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.666271 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.769371 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.769417 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.769427 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.769441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.769465 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.795938 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.795990 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.796003 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.796020 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.796034 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.810714 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:20Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.814472 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.814554 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.814568 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.814585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.814597 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.826735 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:20Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.830648 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.830680 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.830692 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.830709 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.830721 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.844881 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:20Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.848825 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.848879 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.848899 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.848925 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.848952 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.864398 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:20Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.869057 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.869093 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.869102 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.869116 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.869128 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.882375 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:20Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:20 crc kubenswrapper[4990]: E1203 12:38:20.882625 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.884400 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.884445 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.884475 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.884491 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.884501 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.986769 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.986844 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.986853 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.986868 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:20 crc kubenswrapper[4990]: I1203 12:38:20.986877 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:20Z","lastTransitionTime":"2025-12-03T12:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.089177 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.089261 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.089272 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.089289 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.089300 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.191900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.191944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.191954 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.191971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.191989 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.294239 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.294278 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.294289 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.294303 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.294314 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.396506 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.396576 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.396586 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.396602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.396611 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.498610 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.498658 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.498674 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.498696 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.498715 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.601954 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.602042 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.602058 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.602082 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.602099 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.704866 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.705161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.705232 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.705302 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.705365 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.807331 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.807361 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.807370 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.807383 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.807393 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.909260 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.909557 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.909566 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.909579 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:21 crc kubenswrapper[4990]: I1203 12:38:21.909590 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:21Z","lastTransitionTime":"2025-12-03T12:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.012308 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.012353 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.012365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.012380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.012391 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.114982 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.115054 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.115065 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.115081 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.115093 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.217634 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.217675 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.217684 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.217698 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.217709 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.264551 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.264593 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.264663 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:22 crc kubenswrapper[4990]: E1203 12:38:22.264710 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.264737 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:22 crc kubenswrapper[4990]: E1203 12:38:22.264839 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:22 crc kubenswrapper[4990]: E1203 12:38:22.264925 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:22 crc kubenswrapper[4990]: E1203 12:38:22.264979 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.277234 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.290990 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.303711 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.315814 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.319983 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.320031 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.320042 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.320058 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.320068 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.327226 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.337489 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.350993 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.367934 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.379473 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.392646 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.406474 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.419566 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.422095 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.422134 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.422144 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.422157 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.422166 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.431390 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.442862 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.453919 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.468231 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.524179 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.524224 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.524234 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.524251 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.524260 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.627276 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.627976 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.628017 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.628037 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.628053 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.731899 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.731984 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.731999 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.732023 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.732042 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.834234 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.834278 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.834290 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.834309 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.834321 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.936333 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.936380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.936393 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.936409 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:22 crc kubenswrapper[4990]: I1203 12:38:22.936420 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:22Z","lastTransitionTime":"2025-12-03T12:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.039399 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.039483 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.039497 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.039523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.039543 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.142470 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.142511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.142521 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.142538 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.142551 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.244973 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.245209 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.245336 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.245428 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.245552 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.348250 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.348923 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.349054 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.349195 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.349332 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.452388 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.452446 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.452491 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.452518 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.452564 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.555361 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.555442 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.555527 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.555557 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.555577 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.658849 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.658900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.658912 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.658931 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.658945 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.761377 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.761428 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.761439 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.761475 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.761489 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.864420 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.864482 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.864492 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.864508 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.864519 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.967784 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.967840 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.967861 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.967881 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:23 crc kubenswrapper[4990]: I1203 12:38:23.967893 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:23Z","lastTransitionTime":"2025-12-03T12:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.069910 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.069981 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.069994 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.070033 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.070048 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.172624 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.172674 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.172685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.172704 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.172715 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.263748 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.263831 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.263856 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:24 crc kubenswrapper[4990]: E1203 12:38:24.263859 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.263750 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:24 crc kubenswrapper[4990]: E1203 12:38:24.263926 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:24 crc kubenswrapper[4990]: E1203 12:38:24.264068 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:24 crc kubenswrapper[4990]: E1203 12:38:24.264176 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.274370 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.274433 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.274444 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.274476 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.274489 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.376801 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.376844 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.376855 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.376873 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.376884 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.479490 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.479536 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.479547 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.479562 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.479573 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.582000 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.582049 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.582056 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.582079 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.582099 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.685429 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.685611 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.685637 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.685667 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.685691 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.788818 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.789234 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.789439 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.789706 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.789946 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.893161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.893776 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.894015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.894189 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.894325 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.985323 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:24 crc kubenswrapper[4990]: E1203 12:38:24.985529 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:24 crc kubenswrapper[4990]: E1203 12:38:24.985593 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:40.985579314 +0000 UTC m=+69.127490543 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.996838 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.996879 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.996890 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.996903 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:24 crc kubenswrapper[4990]: I1203 12:38:24.996911 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:24Z","lastTransitionTime":"2025-12-03T12:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.100413 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.100471 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.100482 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.100497 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.100507 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.202840 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.202908 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.202923 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.202947 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.202967 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.305251 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.305299 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.305310 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.305328 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.305339 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.410739 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.410790 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.410802 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.410828 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.410841 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.513863 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.513909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.513917 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.513933 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.513944 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.615713 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.615929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.615959 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.615978 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.616000 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.718422 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.718474 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.718485 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.718502 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.718515 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.821124 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.821178 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.821197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.821218 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.821234 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.926811 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.926857 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.926867 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.926887 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:25 crc kubenswrapper[4990]: I1203 12:38:25.926901 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:25Z","lastTransitionTime":"2025-12-03T12:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.029798 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.029864 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.029881 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.029911 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.029937 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.132180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.132231 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.132243 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.132263 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.132275 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.153093 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.165575 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.166754 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.181170 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.194377 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.198844 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.198994 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:38:58.198969668 +0000 UTC m=+86.340880897 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.199031 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.199082 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.199191 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.199221 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.199249 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:58.199237325 +0000 UTC m=+86.341148554 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.199267 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:58.199259075 +0000 UTC m=+86.341170304 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.211340 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.225304 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.234671 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.234697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.234705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.234720 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.234729 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.237679 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.255983 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.263518 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.263562 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.263657 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.263697 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.263732 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.263847 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.263917 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.264018 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.268700 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.280027 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.288959 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.299718 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.299821 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.299940 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.299956 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.299967 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.299997 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.300035 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.300049 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.300011 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:58.299998848 +0000 UTC m=+86.441910077 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:26 crc kubenswrapper[4990]: E1203 12:38:26.300143 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:38:58.300115371 +0000 UTC m=+86.442026600 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.301317 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.314339 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.331728 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.336911 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.336964 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.336975 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.336994 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.337007 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.344601 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.357664 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.370334 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.439442 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.439491 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.439499 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.439512 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.439521 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.542162 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.542207 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.542223 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.542240 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.542252 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.644972 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.645049 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.645061 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.645081 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.645094 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.747607 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.747679 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.747712 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.747732 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.747743 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.849920 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.849962 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.849971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.849988 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.850000 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.952331 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.952376 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.952390 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.952408 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:26 crc kubenswrapper[4990]: I1203 12:38:26.952472 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:26Z","lastTransitionTime":"2025-12-03T12:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.054669 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.054707 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.054715 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.054729 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.054738 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.156594 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.156642 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.156656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.156672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.156684 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.259325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.259387 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.259405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.259429 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.259444 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.264382 4990 scope.go:117] "RemoveContainer" containerID="3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.362260 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.362314 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.362323 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.362339 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.362351 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.464744 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.464821 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.464839 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.464865 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.464882 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.567173 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.567235 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.567251 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.567267 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.567277 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.670086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.670122 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.670132 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.670144 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.670152 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.772408 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.772495 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.772508 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.772522 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.772531 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.875585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.875640 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.875652 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.875670 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.875683 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.978673 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.978707 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.978719 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.978733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:27 crc kubenswrapper[4990]: I1203 12:38:27.978742 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:27Z","lastTransitionTime":"2025-12-03T12:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.081502 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.081600 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.081613 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.081643 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.081660 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.183694 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.183736 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.183746 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.183763 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.183774 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.263851 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.263851 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:28 crc kubenswrapper[4990]: E1203 12:38:28.264023 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.264004 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:28 crc kubenswrapper[4990]: E1203 12:38:28.264115 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.263867 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:28 crc kubenswrapper[4990]: E1203 12:38:28.264299 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:28 crc kubenswrapper[4990]: E1203 12:38:28.264400 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.286358 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.286395 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.286403 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.286415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.286425 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.388682 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.388747 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.388758 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.388772 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.388781 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.490657 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.490698 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.490708 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.490727 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.490738 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.592904 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.592946 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.592955 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.592971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.592981 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.695435 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.695490 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.695501 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.695515 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.695524 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.798549 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.798591 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.798602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.798620 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.798634 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.836438 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/1.log" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.840056 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.840638 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.854044 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.864576 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.878760 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.892613 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.901932 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.901992 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.902006 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.902032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.902047 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:28Z","lastTransitionTime":"2025-12-03T12:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.906057 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.922408 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.939696 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.952750 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.965190 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:28 crc kubenswrapper[4990]: I1203 12:38:28.982442 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.001645 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.004017 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.004069 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.004084 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.004102 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.004119 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.017745 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.039292 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.053668 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.069100 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.084100 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.098420 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.106127 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.106167 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.106177 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.106194 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.106205 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.208937 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.208974 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.208984 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.209001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.209013 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.312146 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.312219 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.312232 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.312250 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.312262 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.415217 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.415283 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.415300 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.415322 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.415335 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.518381 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.518478 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.518493 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.518514 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.518529 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.621539 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.621619 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.621635 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.621655 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.621666 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.725396 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.725494 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.725513 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.725543 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.725571 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.829931 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.829979 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.829989 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.830005 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.830017 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.845134 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/2.log" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.845684 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/1.log" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.848216 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187" exitCode=1 Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.848267 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.848327 4990 scope.go:117] "RemoveContainer" containerID="3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.849118 4990 scope.go:117] "RemoveContainer" containerID="25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187" Dec 03 12:38:29 crc kubenswrapper[4990]: E1203 12:38:29.849427 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.869823 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.883408 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.896636 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.910858 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.923185 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.932875 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.932916 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.932926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.932940 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.932950 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:29Z","lastTransitionTime":"2025-12-03T12:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.939199 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.951947 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.967140 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.983362 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:29 crc kubenswrapper[4990]: I1203 12:38:29.994975 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.007981 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.020676 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.031117 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.034710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.034740 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.034748 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.034760 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.034769 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.042096 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.056855 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.076917 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3835f6344ff24b1e27f31181a902edc2eef7e9504c9857050f685eea644feeb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:12Z\\\",\\\"message\\\":\\\"l_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:38:12.612522 6476 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-marketplace]} name:Service_openshift-marketplace/redhat-marketplace_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.140:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97b6e7b0-06ca-455e-8259-06895040cb0c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 12:38:12.612558 6476 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.092058 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.137686 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.137717 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.137725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.137743 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.137752 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.241303 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.241347 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.241358 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.241374 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.241385 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.264614 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.264637 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.264722 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:30 crc kubenswrapper[4990]: E1203 12:38:30.264738 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.264903 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:30 crc kubenswrapper[4990]: E1203 12:38:30.264997 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:30 crc kubenswrapper[4990]: E1203 12:38:30.265190 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:30 crc kubenswrapper[4990]: E1203 12:38:30.265385 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.344727 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.344779 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.344789 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.344804 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.344815 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.447708 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.447742 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.447755 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.447770 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.447781 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.550604 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.550667 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.550681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.550698 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.550709 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.654376 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.654450 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.654488 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.654517 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.654535 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.757024 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.757081 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.757093 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.757113 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.757125 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.852616 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/2.log" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.855674 4990 scope.go:117] "RemoveContainer" containerID="25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187" Dec 03 12:38:30 crc kubenswrapper[4990]: E1203 12:38:30.855859 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.860368 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.860498 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.860523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.860550 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.860568 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.873762 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.888887 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.909217 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.931015 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.947452 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.963857 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.963904 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.963919 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.963939 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.963952 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:30Z","lastTransitionTime":"2025-12-03T12:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.964130 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.979699 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:30 crc kubenswrapper[4990]: I1203 12:38:30.995638 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.009935 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.025153 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.037222 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.050617 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.065885 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.066869 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.066903 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.066914 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.066939 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.066951 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.082381 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.097661 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.114285 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.126022 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.169680 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.169735 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.169748 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.169766 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.169779 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.180226 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.180275 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.180283 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.180300 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.180313 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: E1203 12:38:31.196975 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.202197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.202263 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.202279 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.202306 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.202323 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: E1203 12:38:31.218021 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.223062 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.223116 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.223125 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.223144 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.223156 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: E1203 12:38:31.238526 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.242422 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.242451 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.242458 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.242472 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.242493 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: E1203 12:38:31.257073 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.262135 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.262171 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.262181 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.262196 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.262210 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: E1203 12:38:31.275423 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:31 crc kubenswrapper[4990]: E1203 12:38:31.275581 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.277122 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.277159 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.277167 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.277180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.277189 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.379980 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.380026 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.380047 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.380066 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.380087 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.482946 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.483004 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.483016 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.483033 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.483043 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.585172 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.585206 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.585216 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.585230 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.585239 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.687115 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.687150 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.687158 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.687169 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.687179 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.789333 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.789382 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.789396 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.789415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.789427 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.892625 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.892667 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.892678 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.892692 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.892702 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.995434 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.995517 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.995529 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.995550 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:31 crc kubenswrapper[4990]: I1203 12:38:31.995564 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:31Z","lastTransitionTime":"2025-12-03T12:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.097472 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.097553 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.097567 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.097585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.097596 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.201622 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.201672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.201682 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.201700 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.201716 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.263712 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.263798 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.263712 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:32 crc kubenswrapper[4990]: E1203 12:38:32.263925 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:32 crc kubenswrapper[4990]: E1203 12:38:32.263880 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.263875 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:32 crc kubenswrapper[4990]: E1203 12:38:32.264294 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:32 crc kubenswrapper[4990]: E1203 12:38:32.264465 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.280862 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.295722 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.304086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.304163 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.304173 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.304192 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.304204 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.310057 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.326390 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.338899 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.354694 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.367942 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.382119 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.395863 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.406430 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.406503 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.406515 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.406532 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.406543 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.409913 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.422589 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.434704 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.447498 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.461924 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.485233 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.500127 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.509009 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.509057 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.509072 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.509089 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.509102 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.514240 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:32Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.611398 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.611434 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.611446 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.611467 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.611496 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.713418 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.713473 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.713483 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.713498 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.713508 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.816116 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.816155 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.816168 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.816185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.816195 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.917912 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.917953 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.917965 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.917980 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:32 crc kubenswrapper[4990]: I1203 12:38:32.917991 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:32Z","lastTransitionTime":"2025-12-03T12:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.020525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.020573 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.020585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.020604 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.020617 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.123324 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.123364 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.123375 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.123387 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.123396 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.225275 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.225578 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.225707 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.225795 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.225859 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.328419 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.328489 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.328501 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.328519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.328530 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.431679 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.431733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.431742 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.431757 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.431768 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.533726 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.533766 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.533777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.533793 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.533804 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.635774 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.635831 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.635841 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.635856 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.635865 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.737868 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.737913 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.737922 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.737947 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.737959 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.840527 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.840596 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.840609 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.840637 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.840650 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.943710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.943756 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.943768 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.943783 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:33 crc kubenswrapper[4990]: I1203 12:38:33.943793 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:33Z","lastTransitionTime":"2025-12-03T12:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.047688 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.047749 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.047768 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.047787 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.047802 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.150161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.150196 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.150204 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.150216 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.150226 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.252350 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.252432 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.252498 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.252524 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.252541 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.264013 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.264053 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:34 crc kubenswrapper[4990]: E1203 12:38:34.264120 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.264214 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.264396 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:34 crc kubenswrapper[4990]: E1203 12:38:34.264403 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:34 crc kubenswrapper[4990]: E1203 12:38:34.264545 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:34 crc kubenswrapper[4990]: E1203 12:38:34.264618 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.355160 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.355209 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.355223 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.355240 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.355253 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.457128 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.457162 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.457171 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.457184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.457194 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.594564 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.594607 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.594618 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.594634 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.594645 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.696935 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.696974 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.696986 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.697001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.697010 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.799805 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.799856 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.799865 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.799884 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.799895 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.902407 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.902441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.902454 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.902487 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:34 crc kubenswrapper[4990]: I1203 12:38:34.902500 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:34Z","lastTransitionTime":"2025-12-03T12:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.005301 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.005349 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.005360 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.005383 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.005397 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.108086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.108130 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.108140 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.108157 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.108168 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.213370 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.213413 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.213423 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.213440 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.213473 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.315650 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.315697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.315756 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.315775 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.315787 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.418098 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.418243 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.418602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.418907 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.418985 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.521779 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.521822 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.521833 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.521850 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.521864 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.625109 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.625157 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.625169 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.625189 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.625202 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.727924 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.727961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.728052 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.728075 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.728086 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.831380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.831422 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.831434 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.831457 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.831498 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.933875 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.933938 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.933952 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.933979 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:35 crc kubenswrapper[4990]: I1203 12:38:35.933995 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:35Z","lastTransitionTime":"2025-12-03T12:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.036528 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.036581 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.036593 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.036613 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.036625 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.139276 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.139306 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.139314 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.139327 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.139335 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.241578 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.241624 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.241635 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.241653 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.241665 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.264504 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.264596 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:36 crc kubenswrapper[4990]: E1203 12:38:36.264661 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.264692 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:36 crc kubenswrapper[4990]: E1203 12:38:36.264739 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.264513 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:36 crc kubenswrapper[4990]: E1203 12:38:36.264859 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:36 crc kubenswrapper[4990]: E1203 12:38:36.264965 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.344045 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.344082 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.344093 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.344113 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.344124 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.446925 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.446972 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.446988 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.447007 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.447020 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.548930 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.548996 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.549007 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.549024 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.549056 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.652714 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.652799 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.652812 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.652828 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.652872 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.756094 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.756151 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.756162 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.756180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.756189 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.858854 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.858891 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.858901 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.858915 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.858940 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.960970 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.961016 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.961028 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.961045 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:36 crc kubenswrapper[4990]: I1203 12:38:36.961055 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:36Z","lastTransitionTime":"2025-12-03T12:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.064102 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.064150 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.064161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.064180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.064192 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.166404 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.166445 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.166477 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.166492 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.166504 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.269060 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.269106 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.269119 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.269136 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.269149 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.371354 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.371430 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.371443 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.371480 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.371495 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.473629 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.473664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.473672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.473687 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.473698 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.576402 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.576688 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.576697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.576712 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.576723 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.679170 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.679489 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.679614 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.679703 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.679785 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.782072 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.782112 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.782123 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.782139 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.782149 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.883872 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.883921 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.883937 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.883954 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.883970 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.985969 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.986271 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.986354 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.986450 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:37 crc kubenswrapper[4990]: I1203 12:38:37.986542 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:37Z","lastTransitionTime":"2025-12-03T12:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.088673 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.088731 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.088744 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.088759 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.088768 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.191259 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.191539 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.191674 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.191767 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.191845 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.263720 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:38 crc kubenswrapper[4990]: E1203 12:38:38.264108 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.264653 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.264758 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:38 crc kubenswrapper[4990]: E1203 12:38:38.264842 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.264690 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:38 crc kubenswrapper[4990]: E1203 12:38:38.265041 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:38 crc kubenswrapper[4990]: E1203 12:38:38.265156 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.294095 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.294374 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.294518 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.294650 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.294963 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.398056 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.398097 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.398106 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.398123 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.398132 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.500555 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.500817 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.500909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.501086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.501186 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.603777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.603842 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.603856 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.603874 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.603885 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.706698 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.706789 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.706807 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.706823 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.706834 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.809223 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.809271 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.809285 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.809302 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.809315 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.911730 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.911830 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.911866 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.911886 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:38 crc kubenswrapper[4990]: I1203 12:38:38.911898 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:38Z","lastTransitionTime":"2025-12-03T12:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.015372 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.015408 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.015419 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.015434 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.015443 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.117930 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.117956 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.117964 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.117978 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.117990 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.221190 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.221221 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.221231 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.221244 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.221254 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.323951 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.323980 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.323990 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.324002 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.324011 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.427180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.427536 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.427630 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.427715 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.427787 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.530737 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.530793 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.530802 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.530867 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.530877 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.633659 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.633693 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.633703 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.633718 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.633728 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.736039 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.736107 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.736121 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.736138 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.736149 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.838410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.839064 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.839160 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.839255 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.839345 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.942187 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.942253 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.942266 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.942283 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:39 crc kubenswrapper[4990]: I1203 12:38:39.942321 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:39Z","lastTransitionTime":"2025-12-03T12:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.045000 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.045044 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.045054 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.045069 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.045083 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.146823 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.146898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.146913 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.146939 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.146953 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.249530 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.249574 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.249584 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.249600 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.249610 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.263702 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.263767 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:40 crc kubenswrapper[4990]: E1203 12:38:40.263837 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:40 crc kubenswrapper[4990]: E1203 12:38:40.263912 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.263973 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:40 crc kubenswrapper[4990]: E1203 12:38:40.264024 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.264175 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:40 crc kubenswrapper[4990]: E1203 12:38:40.264238 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.351475 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.351522 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.351533 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.351548 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.351560 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.454608 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.454675 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.454685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.454701 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.454712 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.557849 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.558138 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.558219 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.558290 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.558369 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.660938 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.661213 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.661293 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.661387 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.661509 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.764320 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.764357 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.764368 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.764383 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.764393 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.866844 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.866889 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.866901 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.866918 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.866930 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.968900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.968946 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.968959 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.968974 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.968985 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:40Z","lastTransitionTime":"2025-12-03T12:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:40 crc kubenswrapper[4990]: I1203 12:38:40.987794 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:40 crc kubenswrapper[4990]: E1203 12:38:40.987965 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:40 crc kubenswrapper[4990]: E1203 12:38:40.988025 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:39:12.988010224 +0000 UTC m=+101.129921453 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.071728 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.071781 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.071792 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.071807 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.071819 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.175367 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.175415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.175426 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.175442 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.175481 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.277482 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.277517 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.277526 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.277539 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.277549 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.315108 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.315156 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.315169 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.315186 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.315197 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: E1203 12:38:41.334842 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:41Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.339915 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.340016 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.340028 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.340051 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.340064 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: E1203 12:38:41.355025 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:41Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.360681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.360723 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.360745 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.360767 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.360778 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: E1203 12:38:41.372781 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:41Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.376239 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.376272 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.376290 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.376307 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.376318 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: E1203 12:38:41.388122 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:41Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.390861 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.390889 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.390898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.390913 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.390924 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: E1203 12:38:41.403201 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:41Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:41 crc kubenswrapper[4990]: E1203 12:38:41.403347 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.404663 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.404696 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.404709 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.404726 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.404737 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.506787 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.506824 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.506834 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.506868 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.506880 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.609352 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.609384 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.609393 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.609408 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.609420 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.712189 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.712264 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.712275 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.712295 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.712308 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.814883 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.815138 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.815237 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.815307 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.815370 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.918471 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.918523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.918535 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.918552 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:41 crc kubenswrapper[4990]: I1203 12:38:41.918568 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:41Z","lastTransitionTime":"2025-12-03T12:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.021060 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.021096 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.021105 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.021119 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.021136 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.123141 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.123178 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.123187 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.123201 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.123211 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.226315 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.226369 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.226380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.226399 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.226441 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.264280 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.264357 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.264413 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:42 crc kubenswrapper[4990]: E1203 12:38:42.264527 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.264610 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:42 crc kubenswrapper[4990]: E1203 12:38:42.264607 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:42 crc kubenswrapper[4990]: E1203 12:38:42.264704 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:42 crc kubenswrapper[4990]: E1203 12:38:42.264772 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.275369 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.290396 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.303358 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.320365 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.328892 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.328927 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.328937 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.328952 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.328963 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.334099 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.347323 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.360080 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.371564 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.388267 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.405404 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.416137 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.429197 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.430783 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.430808 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.430817 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.430829 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.430839 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.441223 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.452876 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.463990 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.478160 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.491568 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:42Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.533581 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.533802 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.533867 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.533960 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.534028 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.636865 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.636906 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.636914 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.636931 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.636942 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.739566 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.739861 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.739951 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.740045 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.740140 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.842955 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.843180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.843378 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.843660 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.843855 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.946746 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.946782 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.946792 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.946825 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:42 crc kubenswrapper[4990]: I1203 12:38:42.946836 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:42Z","lastTransitionTime":"2025-12-03T12:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.049355 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.049405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.049416 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.049431 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.049440 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.151593 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.151645 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.151656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.151677 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.151695 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.255015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.255071 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.255088 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.255106 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.255119 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.357795 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.357838 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.357847 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.357861 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.357869 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.460647 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.460865 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.460926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.460987 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.461053 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.564036 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.564074 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.564083 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.564096 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.564106 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.666357 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.666405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.666415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.666429 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.666439 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.769297 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.769338 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.769349 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.769364 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.769374 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.871916 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.872201 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.872307 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.872413 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.872521 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.975896 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.975957 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.975969 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.975990 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:43 crc kubenswrapper[4990]: I1203 12:38:43.976005 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:43Z","lastTransitionTime":"2025-12-03T12:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.078536 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.078602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.078614 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.078633 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.078647 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.181796 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.181842 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.181851 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.181866 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.181878 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.263769 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:44 crc kubenswrapper[4990]: E1203 12:38:44.263909 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.263785 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.263769 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:44 crc kubenswrapper[4990]: E1203 12:38:44.264399 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:44 crc kubenswrapper[4990]: E1203 12:38:44.264480 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.264735 4990 scope.go:117] "RemoveContainer" containerID="25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.264828 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:44 crc kubenswrapper[4990]: E1203 12:38:44.264899 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:38:44 crc kubenswrapper[4990]: E1203 12:38:44.265034 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.284340 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.284380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.284389 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.284405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.284415 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.386944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.386978 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.386986 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.386998 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.387006 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.489681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.489722 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.489735 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.489750 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.489761 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.592025 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.592063 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.592071 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.592085 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.592094 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.694672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.694725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.694739 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.694755 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.694766 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.797025 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.797055 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.797065 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.797078 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.797087 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.898631 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.898672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.898681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.898695 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:44 crc kubenswrapper[4990]: I1203 12:38:44.898702 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:44Z","lastTransitionTime":"2025-12-03T12:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.000361 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.000885 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.000956 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.001035 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.001108 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.103969 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.104001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.104012 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.104026 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.104036 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.206692 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.206728 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.206737 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.206751 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.206763 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.277129 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.309384 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.309441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.309469 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.309488 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.309503 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.411679 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.411711 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.411720 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.411731 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.411739 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.513515 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.513840 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.513937 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.514023 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.514105 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.616413 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.616472 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.616485 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.616504 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.616514 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.718365 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.718404 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.718415 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.718429 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.718440 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.856961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.857013 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.857027 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.857047 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.857061 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.960260 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.960300 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.960313 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.960330 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:45 crc kubenswrapper[4990]: I1203 12:38:45.960342 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:45Z","lastTransitionTime":"2025-12-03T12:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.062810 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.063480 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.063569 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.063680 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.063748 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.166330 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.166359 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.166368 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.166380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.166389 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.263484 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:46 crc kubenswrapper[4990]: E1203 12:38:46.263620 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.263689 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:46 crc kubenswrapper[4990]: E1203 12:38:46.263758 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.263807 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:46 crc kubenswrapper[4990]: E1203 12:38:46.263861 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.263902 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:46 crc kubenswrapper[4990]: E1203 12:38:46.263948 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.267921 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.267944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.267951 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.267962 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.267970 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.369938 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.369971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.369981 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.369994 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.370005 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.472185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.472223 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.472234 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.472253 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.472264 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.575240 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.575551 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.575629 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.575709 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.575779 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.678968 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.679403 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.679522 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.679617 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.679736 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.783588 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.783654 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.783674 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.783699 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.783716 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.886077 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.886105 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.886113 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.886126 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.886140 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.907132 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/0.log" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.907181 4990 generic.go:334] "Generic (PLEG): container finished" podID="41a42811-350f-4311-bb49-f7f116aa6952" containerID="eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4" exitCode=1 Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.907209 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerDied","Data":"eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.907595 4990 scope.go:117] "RemoveContainer" containerID="eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.922294 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.932312 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.944185 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.958785 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.979054 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.988470 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.988519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.988531 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.988547 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.988559 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:46Z","lastTransitionTime":"2025-12-03T12:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:46 crc kubenswrapper[4990]: I1203 12:38:46.992148 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.008010 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.021925 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.034908 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.050297 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.065045 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"2025-12-03T12:38:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60\\\\n2025-12-03T12:38:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60 to /host/opt/cni/bin/\\\\n2025-12-03T12:38:00Z [verbose] multus-daemon started\\\\n2025-12-03T12:38:00Z [verbose] Readiness Indicator file check\\\\n2025-12-03T12:38:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.078289 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.090924 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"459badf3-3758-45be-b657-2f41cf37ec3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.091293 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.091393 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.091494 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.091592 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.091670 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.102170 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.119149 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.132963 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.147816 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.161602 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.194360 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.194398 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.194410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.194424 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.194436 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.297197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.297230 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.297239 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.297252 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.297261 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.400030 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.400056 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.400063 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.400076 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.400084 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.502070 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.502097 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.502105 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.502117 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.502125 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.604580 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.604623 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.604635 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.604651 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.604665 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.708540 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.708578 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.708587 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.708601 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.708610 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.811575 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.811618 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.811629 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.811646 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.811656 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.919380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.919437 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.919470 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.919493 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.919507 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:47Z","lastTransitionTime":"2025-12-03T12:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.920954 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/0.log" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.921064 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerStarted","Data":"3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea"} Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.938351 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.953898 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.965791 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.979151 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:47 crc kubenswrapper[4990]: I1203 12:38:47.994579 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:47Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.014226 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.022102 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.022143 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.022153 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.022169 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.022188 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.026414 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.037570 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.048645 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.061933 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.071566 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.082565 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"2025-12-03T12:38:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60\\\\n2025-12-03T12:38:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60 to /host/opt/cni/bin/\\\\n2025-12-03T12:38:00Z [verbose] multus-daemon started\\\\n2025-12-03T12:38:00Z [verbose] Readiness Indicator file check\\\\n2025-12-03T12:38:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.094640 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.107189 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"459badf3-3758-45be-b657-2f41cf37ec3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.119612 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.123940 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.123977 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.123987 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.124002 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.124015 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.132425 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.146090 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.158649 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.226701 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.226746 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.226761 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.226779 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.226790 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.263589 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.263672 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.263714 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:48 crc kubenswrapper[4990]: E1203 12:38:48.263734 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.263778 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:48 crc kubenswrapper[4990]: E1203 12:38:48.263890 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:48 crc kubenswrapper[4990]: E1203 12:38:48.263943 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:48 crc kubenswrapper[4990]: E1203 12:38:48.264001 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.328777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.328822 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.328831 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.328863 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.328873 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.431831 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.431890 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.431898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.431914 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.431924 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.535363 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.535398 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.535410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.535433 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.535446 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.638714 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.638765 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.638774 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.638793 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.638803 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.741441 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.741502 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.741511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.741526 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.741536 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.844371 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.844435 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.844631 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.844652 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.844661 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.947476 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.947562 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.947575 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.947592 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:48 crc kubenswrapper[4990]: I1203 12:38:48.947605 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:48Z","lastTransitionTime":"2025-12-03T12:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.049803 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.049846 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.049854 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.049869 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.049879 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.152496 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.152546 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.152561 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.152581 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.152593 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.254682 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.254733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.254752 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.254773 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.254788 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.356640 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.356685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.356698 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.356717 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.356730 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.459182 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.459222 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.459232 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.459247 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.459256 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.561430 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.561508 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.561520 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.561538 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.561549 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.663262 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.663305 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.663317 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.663335 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.663346 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.766572 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.766635 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.766656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.766685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.766699 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.869074 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.869155 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.869169 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.869196 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.869213 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.971246 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.971292 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.971305 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.971321 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:49 crc kubenswrapper[4990]: I1203 12:38:49.971333 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:49Z","lastTransitionTime":"2025-12-03T12:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.074026 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.074380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.074565 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.074705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.074799 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.177465 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.177499 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.177508 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.177521 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.177529 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.264344 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.264370 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.265000 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.265329 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:50 crc kubenswrapper[4990]: E1203 12:38:50.265558 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:50 crc kubenswrapper[4990]: E1203 12:38:50.265419 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:50 crc kubenswrapper[4990]: E1203 12:38:50.265303 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:50 crc kubenswrapper[4990]: E1203 12:38:50.266120 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.279519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.279755 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.279894 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.280047 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.280183 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.382758 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.383186 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.383363 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.383547 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.383676 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.486645 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.486914 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.487105 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.487184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.487259 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.590975 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.591400 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.591513 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.591594 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.591664 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.693901 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.693953 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.693965 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.693979 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.693989 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.796353 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.796395 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.796407 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.796422 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.796432 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.899724 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.899775 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.899788 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.899809 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:50 crc kubenswrapper[4990]: I1203 12:38:50.899824 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:50Z","lastTransitionTime":"2025-12-03T12:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.002074 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.002123 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.002132 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.002147 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.002159 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.104656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.104689 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.104697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.104709 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.104718 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.207086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.207118 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.207125 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.207138 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.207146 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.309398 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.309513 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.309527 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.309548 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.309561 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.412685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.412719 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.412728 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.412742 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.412751 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.490463 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.490491 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.490499 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.490511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.490522 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: E1203 12:38:51.502873 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.508652 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.508835 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.508897 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.508967 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.509038 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: E1203 12:38:51.523213 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.527384 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.527409 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.527436 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.527477 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.527489 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: E1203 12:38:51.538766 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.542544 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.542705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.542781 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.542859 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.542930 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: E1203 12:38:51.556208 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.560372 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.560726 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.560806 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.560906 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.561029 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: E1203 12:38:51.573282 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:51 crc kubenswrapper[4990]: E1203 12:38:51.573689 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.575353 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.575409 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.575425 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.575465 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.575478 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.678088 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.678132 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.678142 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.678156 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.678167 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.780215 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.780525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.780612 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.780689 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.780751 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.883768 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.884099 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.884209 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.884297 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.884601 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.987705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.987956 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.988084 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.988165 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:51 crc kubenswrapper[4990]: I1203 12:38:51.988237 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:51Z","lastTransitionTime":"2025-12-03T12:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.090768 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.090876 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.090885 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.090898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.090906 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.192602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.192852 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.192935 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.193006 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.193081 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.264521 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.264691 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:52 crc kubenswrapper[4990]: E1203 12:38:52.265343 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.264821 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:52 crc kubenswrapper[4990]: E1203 12:38:52.265437 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:52 crc kubenswrapper[4990]: E1203 12:38:52.265270 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.264692 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:52 crc kubenswrapper[4990]: E1203 12:38:52.265546 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.279916 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.292476 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.295334 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.295375 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.295391 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.295412 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.295425 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.302197 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.314858 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.329290 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.350418 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.362295 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.375419 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.391683 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.397074 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.397101 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.397110 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.397124 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.397134 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.404201 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.414465 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.429028 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"2025-12-03T12:38:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60\\\\n2025-12-03T12:38:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60 to /host/opt/cni/bin/\\\\n2025-12-03T12:38:00Z [verbose] multus-daemon started\\\\n2025-12-03T12:38:00Z [verbose] Readiness Indicator file check\\\\n2025-12-03T12:38:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.443379 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.455326 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"459badf3-3758-45be-b657-2f41cf37ec3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.469376 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.482517 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.495084 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.499144 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.499182 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.499191 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.499217 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.499228 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.509005 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.601542 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.601602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.601616 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.601636 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.601649 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.704277 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.704360 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.704372 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.704729 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.704766 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.806951 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.807020 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.807032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.807048 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.807063 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.909306 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.909346 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.909358 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.909376 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:52 crc kubenswrapper[4990]: I1203 12:38:52.909389 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:52Z","lastTransitionTime":"2025-12-03T12:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.012145 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.012184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.012194 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.012212 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.012222 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.114109 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.114150 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.114161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.114176 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.114188 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.216347 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.216382 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.216399 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.216431 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.216441 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.318625 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.318657 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.318666 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.318681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.318691 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.421418 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.421480 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.421494 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.421511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.421525 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.523590 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.523625 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.523634 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.523655 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.523666 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.625468 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.625511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.625525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.625540 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.625550 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.728006 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.728051 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.728062 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.728078 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.728090 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.830633 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.830674 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.830683 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.830697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.830706 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.933337 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.933389 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.933405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.933425 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:53 crc kubenswrapper[4990]: I1203 12:38:53.933441 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:53Z","lastTransitionTime":"2025-12-03T12:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.036278 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.036318 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.036328 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.036344 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.036355 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.138382 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.138426 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.138438 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.138469 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.138484 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.240658 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.240701 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.240710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.240724 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.240734 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.264222 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.264302 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.264353 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.264359 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:54 crc kubenswrapper[4990]: E1203 12:38:54.264487 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:54 crc kubenswrapper[4990]: E1203 12:38:54.264585 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:54 crc kubenswrapper[4990]: E1203 12:38:54.264714 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:54 crc kubenswrapper[4990]: E1203 12:38:54.264831 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.343222 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.343269 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.343281 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.343296 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.343307 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.446047 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.446089 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.446100 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.446116 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.446127 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.548649 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.548681 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.548696 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.548710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.548720 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.650664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.650699 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.650710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.650726 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.650737 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.753437 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.753498 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.753509 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.753525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.753538 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.856339 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.856386 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.856398 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.856416 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.856427 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.958251 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.958286 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.958295 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.958310 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:54 crc kubenswrapper[4990]: I1203 12:38:54.958319 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:54Z","lastTransitionTime":"2025-12-03T12:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.060826 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.060877 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.060891 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.060909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.060923 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.162970 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.163015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.163024 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.163039 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.163048 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.265215 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.265265 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.265276 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.265299 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.265312 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.367857 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.367925 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.367937 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.367957 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.367969 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.470108 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.470171 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.470180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.470203 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.470212 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.572551 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.572791 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.572880 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.572999 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.573092 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.675718 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.675780 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.675790 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.675812 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.675828 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.779479 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.779525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.779538 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.779559 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.779573 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.882963 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.883021 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.883032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.883055 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.883071 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.985803 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.985874 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.985889 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.985923 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:55 crc kubenswrapper[4990]: I1203 12:38:55.985939 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:55Z","lastTransitionTime":"2025-12-03T12:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.088986 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.089028 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.089039 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.089058 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.089068 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.192062 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.192159 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.192176 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.192199 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.192215 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.264232 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.264303 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.264300 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.264232 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:56 crc kubenswrapper[4990]: E1203 12:38:56.264532 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:56 crc kubenswrapper[4990]: E1203 12:38:56.264621 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:56 crc kubenswrapper[4990]: E1203 12:38:56.264739 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:56 crc kubenswrapper[4990]: E1203 12:38:56.264841 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.294602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.294706 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.294729 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.294764 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.294787 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.399153 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.399629 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.399645 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.399666 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.399678 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.502182 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.502530 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.502631 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.502728 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.502816 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.605724 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.605783 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.605793 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.605816 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.605829 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.708959 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.709017 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.709025 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.709041 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.709050 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.812646 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.812723 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.812747 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.812782 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.812810 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.915792 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.916246 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.916335 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.916414 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:56 crc kubenswrapper[4990]: I1203 12:38:56.916506 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:56Z","lastTransitionTime":"2025-12-03T12:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.019240 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.019506 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.019572 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.019649 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.019719 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.122746 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.123001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.123086 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.123161 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.123227 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.225554 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.225584 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.225594 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.225607 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.225616 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.328119 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.328178 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.328191 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.328209 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.328221 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.431656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.431708 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.431720 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.431738 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.431751 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.534111 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.534343 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.534481 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.534557 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.534623 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.637256 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.637340 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.637354 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.637378 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.637393 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.739556 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.739615 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.739625 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.739638 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.739647 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.842585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.842631 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.842642 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.842657 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.842669 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.945051 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.945097 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.945107 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.945123 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:57 crc kubenswrapper[4990]: I1203 12:38:57.945135 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:57Z","lastTransitionTime":"2025-12-03T12:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.047600 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.047641 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.047651 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.047665 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.047675 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.149896 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.149931 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.149941 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.149958 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.149968 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.252410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.252522 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.252540 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.252557 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.252567 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.263801 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.263832 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.263953 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.263949 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.263986 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.264063 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.264201 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.264305 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.265008 4990 scope.go:117] "RemoveContainer" containerID="25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.272578 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.272724 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:02.272706669 +0000 UTC m=+150.414617898 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.272752 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.272783 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.272833 4990 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.272876 4990 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.272901 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:40:02.272883723 +0000 UTC m=+150.414794942 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.272921 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 12:40:02.272912684 +0000 UTC m=+150.414823913 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.354561 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.354590 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.354598 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.354610 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.354619 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.374209 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.374353 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.374839 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.374909 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.374927 4990 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.375010 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 12:40:02.374985535 +0000 UTC m=+150.516896844 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.375008 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.375063 4990 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.375077 4990 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:58 crc kubenswrapper[4990]: E1203 12:38:58.375108 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 12:40:02.375101908 +0000 UTC m=+150.517013137 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.456524 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.457095 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.457165 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.457266 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.457391 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.560041 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.560428 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.560664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.560727 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.560744 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.663844 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.663900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.663912 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.663936 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.663952 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.766888 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.766946 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.766961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.766985 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.767001 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.868750 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.868785 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.868794 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.868808 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.868816 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.975974 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.976018 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.976028 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.976045 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:58 crc kubenswrapper[4990]: I1203 12:38:58.976055 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:58Z","lastTransitionTime":"2025-12-03T12:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.078620 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.078654 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.078662 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.078675 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.078684 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.181821 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.181861 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.181873 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.181896 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.181909 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.284197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.284232 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.284241 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.284254 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.284262 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.386270 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.386546 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.386634 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.386705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.386763 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.488944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.488988 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.489003 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.489023 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.489035 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.590933 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.590961 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.590968 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.590980 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.590989 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.693622 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.693889 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.693994 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.694078 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.694177 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.796184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.796244 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.796259 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.796276 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.796287 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.898357 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.898619 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.898685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.898757 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:38:59 crc kubenswrapper[4990]: I1203 12:38:59.898824 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:38:59Z","lastTransitionTime":"2025-12-03T12:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.000845 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.000901 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.000912 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.000928 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.000937 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.103418 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.103519 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.103538 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.103562 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.103580 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.205337 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.205606 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.205807 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.205988 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.206159 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.263592 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.263667 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:00 crc kubenswrapper[4990]: E1203 12:39:00.263723 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:00 crc kubenswrapper[4990]: E1203 12:39:00.263814 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.263933 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:00 crc kubenswrapper[4990]: E1203 12:39:00.264024 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.264089 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:00 crc kubenswrapper[4990]: E1203 12:39:00.264141 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.308313 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.308362 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.308380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.308402 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.308416 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.410770 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.410823 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.410833 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.410850 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.410863 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.513304 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.513337 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.513349 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.513364 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.513374 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.633707 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.633747 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.633759 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.633779 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.633792 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.736655 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.736688 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.736697 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.736710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.736719 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.838974 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.839013 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.839023 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.839037 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.839047 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.941525 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.941567 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.941577 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.941594 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.941604 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:00Z","lastTransitionTime":"2025-12-03T12:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.959573 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/2.log" Dec 03 12:39:00 crc kubenswrapper[4990]: I1203 12:39:00.962014 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.043523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.043595 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.043607 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.043635 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.043652 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.146479 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.146530 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.146542 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.146561 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.146571 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.249042 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.249089 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.249105 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.249122 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.249132 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.351348 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.351388 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.351400 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.351414 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.351425 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.453652 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.453702 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.453714 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.453733 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.453747 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.556233 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.556288 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.556308 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.556322 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.556331 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.658197 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.658270 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.658285 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.658301 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.658317 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.685395 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.685440 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.685466 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.685489 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.685498 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.697764 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.704679 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.704860 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.704872 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.704888 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.705168 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.719949 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.723620 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.723650 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.723661 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.723675 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.723686 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.737165 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.740786 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.740830 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.740840 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.740856 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.740866 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.752190 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.755471 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.755513 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.755523 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.755538 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.755548 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.766928 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca993a98-c7de-432f-983e-69d030f7ca14\\\",\\\"systemUUID\\\":\\\"a1063038-6a4c-4d75-baa7-5ae3d76716fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.767048 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.768619 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.768658 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.768672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.768688 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.768700 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.870983 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.871023 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.871031 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.871043 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.871052 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.966566 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/3.log" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.967230 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/2.log" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.969593 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" exitCode=1 Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.969627 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.969657 4990 scope.go:117] "RemoveContainer" containerID="25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.970748 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:39:01 crc kubenswrapper[4990]: E1203 12:39:01.971036 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.973334 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.973389 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.973411 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.973439 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.973506 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:01Z","lastTransitionTime":"2025-12-03T12:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:01 crc kubenswrapper[4990]: I1203 12:39:01.986647 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"2025-12-03T12:38:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60\\\\n2025-12-03T12:38:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60 to /host/opt/cni/bin/\\\\n2025-12-03T12:38:00Z [verbose] multus-daemon started\\\\n2025-12-03T12:38:00Z [verbose] Readiness Indicator file check\\\\n2025-12-03T12:38:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.001047 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:01Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.013032 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"459badf3-3758-45be-b657-2f41cf37ec3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.023966 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.039797 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.055322 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.073720 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.076037 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.076094 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.076107 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.076126 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.076138 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.088881 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.108115 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.121063 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.132808 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.148683 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.169199 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:39:01.146372 7087 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:39:01.146470 7087 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-d6l4q\\\\nF1203 12:39:01.146501 7087 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.178128 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.178180 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.178195 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.178214 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.178228 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.181725 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.192901 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.204231 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.216279 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.228442 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.264194 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.264236 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.264194 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:02 crc kubenswrapper[4990]: E1203 12:39:02.264334 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.264359 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:02 crc kubenswrapper[4990]: E1203 12:39:02.264427 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:02 crc kubenswrapper[4990]: E1203 12:39:02.264501 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:02 crc kubenswrapper[4990]: E1203 12:39:02.264743 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.276488 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.280241 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.280273 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.280284 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.280298 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.280307 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.287499 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.300439 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.312263 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.322032 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.332608 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.346084 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.368901 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25a4492a16ceaf5d07f8f1d35ca8d33ad56378dd62819b6af49b1c5238738187\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:29Z\\\",\\\"message\\\":\\\"ube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:38:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 12:38:28.795001 6667 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1203 12:38:28.795009 6667 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1203 12:38:28.794875 6667 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster option\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:39:01.146372 7087 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:39:01.146470 7087 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-d6l4q\\\\nF1203 12:39:01.146501 7087 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.381492 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.382610 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.382654 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.382666 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.382689 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.382701 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.392016 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.403910 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"459badf3-3758-45be-b657-2f41cf37ec3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.414411 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.426574 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"2025-12-03T12:38:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60\\\\n2025-12-03T12:38:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60 to /host/opt/cni/bin/\\\\n2025-12-03T12:38:00Z [verbose] multus-daemon started\\\\n2025-12-03T12:38:00Z [verbose] Readiness Indicator file check\\\\n2025-12-03T12:38:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.437187 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.450712 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.464088 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.476438 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.484518 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.484557 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.484570 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.484586 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.484599 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.490031 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.587129 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.587165 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.587173 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.587185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.587194 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.690202 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.690252 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.690261 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.690279 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.690289 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.791810 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.791841 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.791851 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.791864 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.791873 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.894604 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.894656 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.894667 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.894683 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.894695 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.974310 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/3.log" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.977209 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:39:02 crc kubenswrapper[4990]: E1203 12:39:02.977380 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.991591 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:02Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.996434 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.996513 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.996527 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.996543 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:02 crc kubenswrapper[4990]: I1203 12:39:02.996556 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:02Z","lastTransitionTime":"2025-12-03T12:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.003791 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29fd85428ed73a53e9f825ccff05c8e91a31a490f68d20910cf4e1f3fcbd74e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.014150 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"459badf3-3758-45be-b657-2f41cf37ec3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cce3e44c072af137dddc127f4c12fa7648d0504191652f31bf37fd198902f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bdf24289e4ba14b06cc227ad72896e8092ca09be45b1f5dd08363a5c580bf5ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.023858 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5wptp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaefa7f-af7e-41d7-b729-cb16c06afa2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cccf6de5bc6e7c54021c27295e17c4091f95fc2477705447b037952b4865a4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4kww4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5wptp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.039687 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9c857" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41a42811-350f-4311-bb49-f7f116aa6952\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:38:46Z\\\",\\\"message\\\":\\\"2025-12-03T12:38:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60\\\\n2025-12-03T12:38:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_16be043b-f6d7-4414-99c7-3b432797bc60 to /host/opt/cni/bin/\\\\n2025-12-03T12:38:00Z [verbose] multus-daemon started\\\\n2025-12-03T12:38:00Z [verbose] Readiness Indicator file check\\\\n2025-12-03T12:38:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2mcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9c857\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.052735 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba784e45-d065-42c9-8b4d-f22c277d8e6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be4b38e65270295ee3c7584f27cf80503eb52f396693edda79e8e40d83b3e67b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2fff4427da99319dc262089da85a9953a7a090a838f6b46df5687156e9dbb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtdfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bjsxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.065112 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T12:37:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 12:37:53.582244 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 12:37:53.582443 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 12:37:53.583558 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3160382781/tls.crt::/tmp/serving-cert-3160382781/tls.key\\\\\\\"\\\\nI1203 12:37:53.772741 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 12:37:53.806394 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 12:37:53.806418 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 12:37:53.806440 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 12:37:53.806465 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 12:37:53.810991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 12:37:53.811011 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 12:37:53.811016 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811022 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 12:37:53.811027 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 12:37:53.811030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 12:37:53.811033 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 12:37:53.811036 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 12:37:53.813699 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.077840 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aa1484-90ee-430a-bf93-20f97295a005\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19eda54ace06a2f1433e4e139cc43e604928f668e03718f2f9459e8f0545945b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620e4eaa313206cbbd3d651d27ae46e546cd96c413d5ffd6033b7a9ce6d76d18\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e096d09bc886026f4dbac2e1f25da41802dc9e1684559f2cc3154837d9aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.096022 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.098354 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.098401 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.098410 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.098426 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.098434 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.108180 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db3ba89ff152d369a27d079eb41f255ddf97f9152cae51de1d42a357ad7db8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.120274 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bec8561-741c-4adf-bfce-6f80a0dfb170\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://341bce9a83192adc123042e31bc36a5c3df4fbecb5df0e4e7af5c27bef5f4e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9z2gt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-85qrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.134797 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2kt75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc81041d-8c71-4fcb-8901-8047795b21c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20daee71880fd68e59826e5978ca281746775522b5bf3b52caae5a6b18615055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bf8a455223680830e1c97762eaa92140114ff7247f5adee08fb4d0c84462662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115fc3d905b48a06d7aea3020bf6f4ebbc30d21c616c47ebe4ace30d3a09f278\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3d7c87e3c572423c53eecd48ebc69b627174f6061dfd90271fe489e1302a89d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6237e4d8087b60e1c48329faf1f8ba42e2e29454a37a5084926e3329777c78cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6784683b86044f5013702227f5f6742953fc684c0ac78b5ab53d0b61536503f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f5d4fa23d63b1cd40a2afabed9826022e0125ec9f3f610a39fc773c8860652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:38:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llqxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2kt75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.152543 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T12:39:01Z\\\",\\\"message\\\":\\\"48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:39:01.146372 7087 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 12:39:01.146470 7087 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-d6l4q\\\\nF1203 12:39:01.146501 7087 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T12:39:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42qbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7lrq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.162325 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a061fb3f-ace6-4ab6-a499-b34979186d08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zb9nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:38:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6sqk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.172815 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71deade3-11e0-49f8-ae61-30ad5577a6c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:38:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca556a95a23598d2b873625e0e3ce3d994bb4e369150f453fa53fce73ca24b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2a880fb0890feaf71fa33414652460b9bd3e6e4a065c47bead4c76cb70259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae8c241115f3d5b5290a3b98056c4ad6949c7bfb87cbd5dc8ea39d8b02b2240d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38cad1f47ef027a0c1144512e5e71134e4b6d39c52e28d3c0dd55b696698309d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T12:37:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T12:37:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.185651 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.197312 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077e872963784d09c7eb8eb8acea815212daeb90fb3523bff0aeafd8931a01b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60c35cb08ffe3c80fe7d49bffb573b2db080a2ae36ae4383f77f88ce1473ee7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.200866 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.200889 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.200897 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.200909 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.200917 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.207387 4990 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d6l4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1144ccf4-9fd2-4705-9e8c-0ddace5364fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T12:37:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://973954c64888fd6a8c8463696343ca366597a2031316acc3fff49da8cd83a791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T12:37:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pkvlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T12:37:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d6l4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T12:39:03Z is after 2025-08-24T17:21:41Z" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.303380 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.303419 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.303429 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.303444 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.303470 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.405921 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.405966 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.405978 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.405997 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.406011 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.507873 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.507919 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.507928 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.507942 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.507952 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.610848 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.610898 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.610908 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.610929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.610941 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.713679 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.713765 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.713784 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.713816 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.713838 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.817408 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.817711 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.817727 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.817753 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.817769 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.920718 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.920784 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.920798 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.920815 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:03 crc kubenswrapper[4990]: I1203 12:39:03.920826 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:03Z","lastTransitionTime":"2025-12-03T12:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.023536 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.023585 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.023599 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.023616 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.023629 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.126071 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.126103 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.126112 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.126125 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.126133 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.228296 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.228347 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.228356 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.228371 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.228381 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.263680 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.263707 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.263712 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:04 crc kubenswrapper[4990]: E1203 12:39:04.263856 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.263894 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:04 crc kubenswrapper[4990]: E1203 12:39:04.263993 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:04 crc kubenswrapper[4990]: E1203 12:39:04.264085 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:04 crc kubenswrapper[4990]: E1203 12:39:04.264183 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.330766 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.330859 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.330871 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.330889 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.330903 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.433632 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.433921 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.434062 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.434208 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.434388 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.537635 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.537929 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.538010 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.538091 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.538176 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.640526 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.640571 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.640584 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.640602 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.640613 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.742846 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.742886 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.742897 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.742913 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.742925 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.849651 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.849703 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.849715 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.849732 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.849741 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.951926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.952551 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.952777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.952983 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:04 crc kubenswrapper[4990]: I1203 12:39:04.953131 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:04Z","lastTransitionTime":"2025-12-03T12:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.055725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.055764 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.055773 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.055788 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.055799 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.158013 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.158052 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.158061 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.158075 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.158085 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.259974 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.260022 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.260033 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.260049 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.260059 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.362137 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.362184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.362206 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.362225 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.362237 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.464024 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.464069 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.464080 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.464094 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.464106 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.566284 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.566317 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.566326 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.566342 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.566353 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.669298 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.669334 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.669344 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.669359 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.669371 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.771422 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.771529 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.771545 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.771561 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.771570 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.874423 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.874504 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.874520 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.874542 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.874555 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.976751 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.976786 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.976796 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.976810 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:05 crc kubenswrapper[4990]: I1203 12:39:05.976819 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:05Z","lastTransitionTime":"2025-12-03T12:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.078995 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.079032 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.079041 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.079057 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.079066 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.181292 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.181325 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.181333 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.181345 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.181354 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.263793 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.263864 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:06 crc kubenswrapper[4990]: E1203 12:39:06.263900 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:06 crc kubenswrapper[4990]: E1203 12:39:06.263994 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.264053 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:06 crc kubenswrapper[4990]: E1203 12:39:06.264096 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.264225 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:06 crc kubenswrapper[4990]: E1203 12:39:06.264280 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.283657 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.283725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.283738 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.283755 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.283768 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.386184 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.386218 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.386227 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.386240 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.386250 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.489405 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.489531 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.489555 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.489581 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.489603 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.592155 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.592185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.592193 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.592205 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.592213 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.694175 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.694221 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.694231 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.694246 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.694258 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.796101 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.796132 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.796141 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.796155 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.796164 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.898857 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.898903 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.898916 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.898932 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:06 crc kubenswrapper[4990]: I1203 12:39:06.898944 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:06Z","lastTransitionTime":"2025-12-03T12:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.001171 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.001200 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.001210 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.001223 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.001234 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.103989 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.104039 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.104052 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.104106 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.104171 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.206748 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.206786 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.206798 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.206816 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.206830 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.308926 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.308971 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.308983 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.308999 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.309011 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.411334 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.411390 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.411403 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.411421 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.411432 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.514075 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.514115 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.514126 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.514141 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.514152 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.616498 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.616527 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.616536 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.616549 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.616560 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.718690 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.718732 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.718744 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.718760 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.718773 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.821140 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.821185 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.821196 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.821213 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.821231 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.924712 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.924874 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.924900 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.924918 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:07 crc kubenswrapper[4990]: I1203 12:39:07.924931 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:07Z","lastTransitionTime":"2025-12-03T12:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.027297 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.027345 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.027359 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.027374 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.027385 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.131298 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.131344 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.131356 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.131372 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.131381 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.233676 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.233721 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.233735 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.233752 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.233762 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.263914 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.264004 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:08 crc kubenswrapper[4990]: E1203 12:39:08.264045 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.264085 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.264116 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:08 crc kubenswrapper[4990]: E1203 12:39:08.264165 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:08 crc kubenswrapper[4990]: E1203 12:39:08.264221 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:08 crc kubenswrapper[4990]: E1203 12:39:08.264297 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.337015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.337064 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.337073 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.337090 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.337102 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.439435 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.439499 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.439511 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.439529 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.439541 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.542375 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.542438 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.542469 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.542487 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.542500 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.644741 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.644776 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.644783 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.644798 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.644806 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.746778 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.747316 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.747440 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.747565 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.747672 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.850482 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.850827 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.850916 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.851001 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.851099 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.954188 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.954252 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.954262 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.954275 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:08 crc kubenswrapper[4990]: I1203 12:39:08.954283 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:08Z","lastTransitionTime":"2025-12-03T12:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.057015 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.057101 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.057179 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.057199 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.057211 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.160071 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.160651 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.160685 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.160704 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.160715 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.263079 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.263128 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.263142 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.263159 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.263188 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.365191 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.365228 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.365238 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.365254 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.365265 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.467578 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.467615 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.467623 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.467636 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.467646 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.569705 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.569760 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.569777 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.569798 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.569813 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.672394 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.672670 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.672738 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.672820 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.672904 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.775740 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.776626 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.776730 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.776867 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.777001 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.878559 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.878588 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.878598 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.878609 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.878617 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.980687 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.980725 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.980734 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.980746 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:09 crc kubenswrapper[4990]: I1203 12:39:09.980755 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:09Z","lastTransitionTime":"2025-12-03T12:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.082876 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.082922 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.082936 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.082953 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.082969 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.185895 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.186224 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.186292 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.186363 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.186438 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.263733 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.263822 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.263741 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:10 crc kubenswrapper[4990]: E1203 12:39:10.263869 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:10 crc kubenswrapper[4990]: E1203 12:39:10.263969 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:10 crc kubenswrapper[4990]: E1203 12:39:10.264101 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.264268 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:10 crc kubenswrapper[4990]: E1203 12:39:10.264470 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.288866 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.288908 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.288927 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.288944 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.288953 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.390636 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.390672 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.390684 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.390699 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.390709 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.493250 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.493301 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.493312 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.493329 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.493340 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.595764 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.595801 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.595810 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.595822 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.595831 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.697921 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.697954 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.697964 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.697977 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.697986 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.800260 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.800303 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.800315 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.800331 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.800343 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.903186 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.903257 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.903271 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.903289 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:10 crc kubenswrapper[4990]: I1203 12:39:10.903300 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:10Z","lastTransitionTime":"2025-12-03T12:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.005758 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.005791 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.005799 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.005811 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.005819 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.107996 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.108052 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.108063 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.108080 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.108091 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.210194 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.210250 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.210263 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.210283 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.210300 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.276438 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.312235 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.312479 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.312571 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.312738 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.312857 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.415324 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.415371 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.415384 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.415402 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.415413 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.517737 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.517789 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.517807 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.517829 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.517849 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.621254 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.621293 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.621302 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.621318 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.621329 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.723664 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.723710 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.723719 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.723741 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.723754 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.826802 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.826851 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.826863 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.826880 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.826893 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.929841 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.929908 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.929919 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.929935 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:11 crc kubenswrapper[4990]: I1203 12:39:11.929971 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:11Z","lastTransitionTime":"2025-12-03T12:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.032111 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.032160 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.032171 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.032190 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.032204 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:12Z","lastTransitionTime":"2025-12-03T12:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.125579 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.125617 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.125625 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.125640 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.125650 4990 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T12:39:12Z","lastTransitionTime":"2025-12-03T12:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.168440 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92"] Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.168934 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.170949 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.173134 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.173584 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.173869 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.203986 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.203962099 podStartE2EDuration="1m18.203962099s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.188990987 +0000 UTC m=+100.330902216" watchObservedRunningTime="2025-12-03 12:39:12.203962099 +0000 UTC m=+100.345873328" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.219147 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=74.219126286 podStartE2EDuration="1m14.219126286s" podCreationTimestamp="2025-12-03 12:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.20441757 +0000 UTC m=+100.346328799" watchObservedRunningTime="2025-12-03 12:39:12.219126286 +0000 UTC m=+100.361037515" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.263757 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:12 crc kubenswrapper[4990]: E1203 12:39:12.263855 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.263918 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:12 crc kubenswrapper[4990]: E1203 12:39:12.263965 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.264187 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:12 crc kubenswrapper[4990]: E1203 12:39:12.264237 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.264417 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:12 crc kubenswrapper[4990]: E1203 12:39:12.264492 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.300113 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.300092849 podStartE2EDuration="46.300092849s" podCreationTimestamp="2025-12-03 12:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.285704322 +0000 UTC m=+100.427615551" watchObservedRunningTime="2025-12-03 12:39:12.300092849 +0000 UTC m=+100.442004078" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.320169 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/06d2e904-5f71-4f52-9c58-6f1712dba0a7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.320250 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/06d2e904-5f71-4f52-9c58-6f1712dba0a7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.320344 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06d2e904-5f71-4f52-9c58-6f1712dba0a7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.320384 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06d2e904-5f71-4f52-9c58-6f1712dba0a7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.320412 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06d2e904-5f71-4f52-9c58-6f1712dba0a7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.349505 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podStartSLOduration=78.349473267 podStartE2EDuration="1m18.349473267s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.348720238 +0000 UTC m=+100.490631477" watchObservedRunningTime="2025-12-03 12:39:12.349473267 +0000 UTC m=+100.491384506" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.350051 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-d6l4q" podStartSLOduration=78.350043641 podStartE2EDuration="1m18.350043641s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.337172413 +0000 UTC m=+100.479083662" watchObservedRunningTime="2025-12-03 12:39:12.350043641 +0000 UTC m=+100.491954870" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.368267 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2kt75" podStartSLOduration=78.368247875 podStartE2EDuration="1m18.368247875s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.367791384 +0000 UTC m=+100.509702623" watchObservedRunningTime="2025-12-03 12:39:12.368247875 +0000 UTC m=+100.510159104" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.418292 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.41827254 podStartE2EDuration="1.41827254s" podCreationTimestamp="2025-12-03 12:39:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.417463029 +0000 UTC m=+100.559374278" watchObservedRunningTime="2025-12-03 12:39:12.41827254 +0000 UTC m=+100.560183769" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421137 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06d2e904-5f71-4f52-9c58-6f1712dba0a7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421181 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06d2e904-5f71-4f52-9c58-6f1712dba0a7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421237 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/06d2e904-5f71-4f52-9c58-6f1712dba0a7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421302 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/06d2e904-5f71-4f52-9c58-6f1712dba0a7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421369 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06d2e904-5f71-4f52-9c58-6f1712dba0a7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421444 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/06d2e904-5f71-4f52-9c58-6f1712dba0a7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.421482 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/06d2e904-5f71-4f52-9c58-6f1712dba0a7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.422146 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06d2e904-5f71-4f52-9c58-6f1712dba0a7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.429188 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06d2e904-5f71-4f52-9c58-6f1712dba0a7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.439845 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06d2e904-5f71-4f52-9c58-6f1712dba0a7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dbg92\" (UID: \"06d2e904-5f71-4f52-9c58-6f1712dba0a7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.483989 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=27.483972135 podStartE2EDuration="27.483972135s" podCreationTimestamp="2025-12-03 12:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.461800669 +0000 UTC m=+100.603711898" watchObservedRunningTime="2025-12-03 12:39:12.483972135 +0000 UTC m=+100.625883364" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.489395 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.498415 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5wptp" podStartSLOduration=78.49832741 podStartE2EDuration="1m18.49832741s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.483507383 +0000 UTC m=+100.625418612" watchObservedRunningTime="2025-12-03 12:39:12.49832741 +0000 UTC m=+100.640238639" Dec 03 12:39:12 crc kubenswrapper[4990]: I1203 12:39:12.516757 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-9c857" podStartSLOduration=78.516733999 podStartE2EDuration="1m18.516733999s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.498732111 +0000 UTC m=+100.640643340" watchObservedRunningTime="2025-12-03 12:39:12.516733999 +0000 UTC m=+100.658645228" Dec 03 12:39:13 crc kubenswrapper[4990]: I1203 12:39:13.007852 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" event={"ID":"06d2e904-5f71-4f52-9c58-6f1712dba0a7","Type":"ContainerStarted","Data":"413461d64dbf63deee7cf2139fdf899e608f924f0ac2fbdb291654a75b5f79c2"} Dec 03 12:39:13 crc kubenswrapper[4990]: I1203 12:39:13.030280 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:13 crc kubenswrapper[4990]: E1203 12:39:13.030421 4990 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:39:13 crc kubenswrapper[4990]: E1203 12:39:13.030537 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs podName:a061fb3f-ace6-4ab6-a499-b34979186d08 nodeName:}" failed. No retries permitted until 2025-12-03 12:40:17.030519352 +0000 UTC m=+165.172430571 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs") pod "network-metrics-daemon-6sqk8" (UID: "a061fb3f-ace6-4ab6-a499-b34979186d08") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 12:39:13 crc kubenswrapper[4990]: I1203 12:39:13.548544 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:39:13 crc kubenswrapper[4990]: I1203 12:39:13.549712 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:39:13 crc kubenswrapper[4990]: E1203 12:39:13.549928 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.011265 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" event={"ID":"06d2e904-5f71-4f52-9c58-6f1712dba0a7","Type":"ContainerStarted","Data":"ad363e5bc415fadefae94c25e6c51d5acd8e0dfc48fde0740476d04f08452322"} Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.025490 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bjsxk" podStartSLOduration=79.025473996 podStartE2EDuration="1m19.025473996s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:12.517051108 +0000 UTC m=+100.658962337" watchObservedRunningTime="2025-12-03 12:39:14.025473996 +0000 UTC m=+102.167385215" Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.025904 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dbg92" podStartSLOduration=80.025898357 podStartE2EDuration="1m20.025898357s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:14.025269311 +0000 UTC m=+102.167180540" watchObservedRunningTime="2025-12-03 12:39:14.025898357 +0000 UTC m=+102.167809596" Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.263580 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.263615 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.264185 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:14 crc kubenswrapper[4990]: I1203 12:39:14.264263 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:14 crc kubenswrapper[4990]: E1203 12:39:14.264268 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:14 crc kubenswrapper[4990]: E1203 12:39:14.264357 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:14 crc kubenswrapper[4990]: E1203 12:39:14.264490 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:14 crc kubenswrapper[4990]: E1203 12:39:14.265075 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:16 crc kubenswrapper[4990]: I1203 12:39:16.264432 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:16 crc kubenswrapper[4990]: I1203 12:39:16.264504 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:16 crc kubenswrapper[4990]: E1203 12:39:16.264563 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:16 crc kubenswrapper[4990]: I1203 12:39:16.264642 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:16 crc kubenswrapper[4990]: E1203 12:39:16.264717 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:16 crc kubenswrapper[4990]: I1203 12:39:16.264772 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:16 crc kubenswrapper[4990]: E1203 12:39:16.264826 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:16 crc kubenswrapper[4990]: E1203 12:39:16.264996 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:18 crc kubenswrapper[4990]: I1203 12:39:18.263604 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:18 crc kubenswrapper[4990]: E1203 12:39:18.263747 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:18 crc kubenswrapper[4990]: I1203 12:39:18.264321 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:18 crc kubenswrapper[4990]: I1203 12:39:18.264430 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:18 crc kubenswrapper[4990]: E1203 12:39:18.264528 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:18 crc kubenswrapper[4990]: I1203 12:39:18.264412 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:18 crc kubenswrapper[4990]: E1203 12:39:18.264804 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:18 crc kubenswrapper[4990]: E1203 12:39:18.264883 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:20 crc kubenswrapper[4990]: I1203 12:39:20.264243 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:20 crc kubenswrapper[4990]: I1203 12:39:20.264271 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:20 crc kubenswrapper[4990]: E1203 12:39:20.264331 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:20 crc kubenswrapper[4990]: E1203 12:39:20.264421 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:20 crc kubenswrapper[4990]: I1203 12:39:20.264425 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:20 crc kubenswrapper[4990]: I1203 12:39:20.264243 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:20 crc kubenswrapper[4990]: E1203 12:39:20.264621 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:20 crc kubenswrapper[4990]: E1203 12:39:20.264715 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:22 crc kubenswrapper[4990]: I1203 12:39:22.264256 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:22 crc kubenswrapper[4990]: I1203 12:39:22.264218 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:22 crc kubenswrapper[4990]: I1203 12:39:22.265513 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:22 crc kubenswrapper[4990]: E1203 12:39:22.265363 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:22 crc kubenswrapper[4990]: I1203 12:39:22.265419 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:22 crc kubenswrapper[4990]: E1203 12:39:22.265680 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:22 crc kubenswrapper[4990]: E1203 12:39:22.265723 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:22 crc kubenswrapper[4990]: E1203 12:39:22.265849 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:24 crc kubenswrapper[4990]: I1203 12:39:24.263714 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:24 crc kubenswrapper[4990]: E1203 12:39:24.264086 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:24 crc kubenswrapper[4990]: I1203 12:39:24.263822 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:24 crc kubenswrapper[4990]: I1203 12:39:24.263737 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:24 crc kubenswrapper[4990]: E1203 12:39:24.264153 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:24 crc kubenswrapper[4990]: I1203 12:39:24.263868 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:24 crc kubenswrapper[4990]: E1203 12:39:24.264276 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:24 crc kubenswrapper[4990]: E1203 12:39:24.264350 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:26 crc kubenswrapper[4990]: I1203 12:39:26.264045 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:26 crc kubenswrapper[4990]: I1203 12:39:26.264086 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:26 crc kubenswrapper[4990]: I1203 12:39:26.264201 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:26 crc kubenswrapper[4990]: E1203 12:39:26.264199 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:26 crc kubenswrapper[4990]: E1203 12:39:26.264283 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:26 crc kubenswrapper[4990]: I1203 12:39:26.264317 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:26 crc kubenswrapper[4990]: E1203 12:39:26.264362 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:26 crc kubenswrapper[4990]: E1203 12:39:26.264431 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:28 crc kubenswrapper[4990]: I1203 12:39:28.264318 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:28 crc kubenswrapper[4990]: I1203 12:39:28.264386 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:28 crc kubenswrapper[4990]: E1203 12:39:28.264484 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:28 crc kubenswrapper[4990]: I1203 12:39:28.264508 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:28 crc kubenswrapper[4990]: E1203 12:39:28.264608 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:28 crc kubenswrapper[4990]: I1203 12:39:28.264652 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:28 crc kubenswrapper[4990]: E1203 12:39:28.264939 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:28 crc kubenswrapper[4990]: E1203 12:39:28.265034 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:28 crc kubenswrapper[4990]: I1203 12:39:28.265282 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:39:28 crc kubenswrapper[4990]: E1203 12:39:28.265479 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:39:30 crc kubenswrapper[4990]: I1203 12:39:30.264014 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:30 crc kubenswrapper[4990]: I1203 12:39:30.264078 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:30 crc kubenswrapper[4990]: I1203 12:39:30.264055 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:30 crc kubenswrapper[4990]: I1203 12:39:30.264244 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:30 crc kubenswrapper[4990]: E1203 12:39:30.264234 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:30 crc kubenswrapper[4990]: E1203 12:39:30.264303 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:30 crc kubenswrapper[4990]: E1203 12:39:30.264347 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:30 crc kubenswrapper[4990]: E1203 12:39:30.264547 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:32 crc kubenswrapper[4990]: I1203 12:39:32.264185 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:32 crc kubenswrapper[4990]: I1203 12:39:32.264185 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:32 crc kubenswrapper[4990]: I1203 12:39:32.264238 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:32 crc kubenswrapper[4990]: I1203 12:39:32.264296 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:32 crc kubenswrapper[4990]: E1203 12:39:32.267533 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:32 crc kubenswrapper[4990]: E1203 12:39:32.267893 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:32 crc kubenswrapper[4990]: E1203 12:39:32.268123 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:32 crc kubenswrapper[4990]: E1203 12:39:32.268337 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:32 crc kubenswrapper[4990]: E1203 12:39:32.282675 4990 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 12:39:33 crc kubenswrapper[4990]: I1203 12:39:33.075689 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/1.log" Dec 03 12:39:33 crc kubenswrapper[4990]: I1203 12:39:33.076082 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/0.log" Dec 03 12:39:33 crc kubenswrapper[4990]: I1203 12:39:33.076126 4990 generic.go:334] "Generic (PLEG): container finished" podID="41a42811-350f-4311-bb49-f7f116aa6952" containerID="3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea" exitCode=1 Dec 03 12:39:33 crc kubenswrapper[4990]: I1203 12:39:33.076164 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerDied","Data":"3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea"} Dec 03 12:39:33 crc kubenswrapper[4990]: I1203 12:39:33.076214 4990 scope.go:117] "RemoveContainer" containerID="eb2dff35d1cc6063fa436cce04cd10a55c4e87e9eab7450a7930a95c041d5ec4" Dec 03 12:39:33 crc kubenswrapper[4990]: I1203 12:39:33.076732 4990 scope.go:117] "RemoveContainer" containerID="3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea" Dec 03 12:39:33 crc kubenswrapper[4990]: E1203 12:39:33.076951 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-9c857_openshift-multus(41a42811-350f-4311-bb49-f7f116aa6952)\"" pod="openshift-multus/multus-9c857" podUID="41a42811-350f-4311-bb49-f7f116aa6952" Dec 03 12:39:33 crc kubenswrapper[4990]: E1203 12:39:33.085765 4990 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 12:39:34 crc kubenswrapper[4990]: I1203 12:39:34.081846 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/1.log" Dec 03 12:39:34 crc kubenswrapper[4990]: I1203 12:39:34.264307 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:34 crc kubenswrapper[4990]: I1203 12:39:34.264402 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:34 crc kubenswrapper[4990]: E1203 12:39:34.264475 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:34 crc kubenswrapper[4990]: I1203 12:39:34.264307 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:34 crc kubenswrapper[4990]: E1203 12:39:34.264627 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:34 crc kubenswrapper[4990]: I1203 12:39:34.264328 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:34 crc kubenswrapper[4990]: E1203 12:39:34.264807 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:34 crc kubenswrapper[4990]: E1203 12:39:34.264973 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:36 crc kubenswrapper[4990]: I1203 12:39:36.263577 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:36 crc kubenswrapper[4990]: I1203 12:39:36.263604 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:36 crc kubenswrapper[4990]: E1203 12:39:36.263696 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:36 crc kubenswrapper[4990]: I1203 12:39:36.263735 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:36 crc kubenswrapper[4990]: E1203 12:39:36.263827 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:36 crc kubenswrapper[4990]: I1203 12:39:36.263832 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:36 crc kubenswrapper[4990]: E1203 12:39:36.263894 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:36 crc kubenswrapper[4990]: E1203 12:39:36.264029 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:38 crc kubenswrapper[4990]: E1203 12:39:38.086975 4990 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 12:39:38 crc kubenswrapper[4990]: I1203 12:39:38.264565 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:38 crc kubenswrapper[4990]: I1203 12:39:38.264590 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:38 crc kubenswrapper[4990]: E1203 12:39:38.264731 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:38 crc kubenswrapper[4990]: I1203 12:39:38.264761 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:38 crc kubenswrapper[4990]: I1203 12:39:38.264590 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:38 crc kubenswrapper[4990]: E1203 12:39:38.264889 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:38 crc kubenswrapper[4990]: E1203 12:39:38.264984 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:38 crc kubenswrapper[4990]: E1203 12:39:38.265055 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:40 crc kubenswrapper[4990]: I1203 12:39:40.264079 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:40 crc kubenswrapper[4990]: I1203 12:39:40.264104 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:40 crc kubenswrapper[4990]: I1203 12:39:40.264129 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:40 crc kubenswrapper[4990]: I1203 12:39:40.264184 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:40 crc kubenswrapper[4990]: E1203 12:39:40.264402 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:40 crc kubenswrapper[4990]: E1203 12:39:40.264576 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:40 crc kubenswrapper[4990]: E1203 12:39:40.264794 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:40 crc kubenswrapper[4990]: E1203 12:39:40.264842 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:40 crc kubenswrapper[4990]: I1203 12:39:40.265803 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:39:40 crc kubenswrapper[4990]: E1203 12:39:40.266146 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7lrq6_openshift-ovn-kubernetes(f5d4bd47-2fa6-4e6a-9697-9af4e227afe1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" Dec 03 12:39:42 crc kubenswrapper[4990]: I1203 12:39:42.264295 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:42 crc kubenswrapper[4990]: I1203 12:39:42.264487 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:42 crc kubenswrapper[4990]: I1203 12:39:42.264491 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:42 crc kubenswrapper[4990]: I1203 12:39:42.264532 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:42 crc kubenswrapper[4990]: E1203 12:39:42.265547 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:42 crc kubenswrapper[4990]: E1203 12:39:42.265733 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:42 crc kubenswrapper[4990]: E1203 12:39:42.265834 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:42 crc kubenswrapper[4990]: E1203 12:39:42.265872 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:43 crc kubenswrapper[4990]: E1203 12:39:43.087514 4990 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 12:39:44 crc kubenswrapper[4990]: I1203 12:39:44.263907 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:44 crc kubenswrapper[4990]: E1203 12:39:44.264041 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:44 crc kubenswrapper[4990]: I1203 12:39:44.264217 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:44 crc kubenswrapper[4990]: E1203 12:39:44.264266 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:44 crc kubenswrapper[4990]: I1203 12:39:44.264417 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:44 crc kubenswrapper[4990]: E1203 12:39:44.264571 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:44 crc kubenswrapper[4990]: I1203 12:39:44.264585 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:44 crc kubenswrapper[4990]: E1203 12:39:44.264796 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:45 crc kubenswrapper[4990]: I1203 12:39:45.263889 4990 scope.go:117] "RemoveContainer" containerID="3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea" Dec 03 12:39:46 crc kubenswrapper[4990]: I1203 12:39:46.117367 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/1.log" Dec 03 12:39:46 crc kubenswrapper[4990]: I1203 12:39:46.117707 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerStarted","Data":"82d1c99f98ceb0301fbb2c407c935e3b0fd5c83937c77e454284191b586dc10c"} Dec 03 12:39:46 crc kubenswrapper[4990]: I1203 12:39:46.264098 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:46 crc kubenswrapper[4990]: I1203 12:39:46.264146 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:46 crc kubenswrapper[4990]: I1203 12:39:46.264167 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:46 crc kubenswrapper[4990]: E1203 12:39:46.264229 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:46 crc kubenswrapper[4990]: I1203 12:39:46.264259 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:46 crc kubenswrapper[4990]: E1203 12:39:46.264393 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:46 crc kubenswrapper[4990]: E1203 12:39:46.264428 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:46 crc kubenswrapper[4990]: E1203 12:39:46.264505 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:48 crc kubenswrapper[4990]: E1203 12:39:48.088524 4990 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 12:39:48 crc kubenswrapper[4990]: I1203 12:39:48.263523 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:48 crc kubenswrapper[4990]: I1203 12:39:48.263541 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:48 crc kubenswrapper[4990]: I1203 12:39:48.263576 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:48 crc kubenswrapper[4990]: E1203 12:39:48.263687 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:48 crc kubenswrapper[4990]: E1203 12:39:48.263773 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:48 crc kubenswrapper[4990]: I1203 12:39:48.263841 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:48 crc kubenswrapper[4990]: E1203 12:39:48.263845 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:48 crc kubenswrapper[4990]: E1203 12:39:48.264042 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:50 crc kubenswrapper[4990]: I1203 12:39:50.263530 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:50 crc kubenswrapper[4990]: I1203 12:39:50.263548 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:50 crc kubenswrapper[4990]: E1203 12:39:50.264381 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:50 crc kubenswrapper[4990]: I1203 12:39:50.263660 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:50 crc kubenswrapper[4990]: E1203 12:39:50.264545 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:50 crc kubenswrapper[4990]: I1203 12:39:50.263582 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:50 crc kubenswrapper[4990]: E1203 12:39:50.264772 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:50 crc kubenswrapper[4990]: E1203 12:39:50.265316 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:52 crc kubenswrapper[4990]: I1203 12:39:52.264425 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:52 crc kubenswrapper[4990]: I1203 12:39:52.264528 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:52 crc kubenswrapper[4990]: I1203 12:39:52.264426 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:52 crc kubenswrapper[4990]: I1203 12:39:52.264565 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:52 crc kubenswrapper[4990]: E1203 12:39:52.265896 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:52 crc kubenswrapper[4990]: E1203 12:39:52.266048 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:52 crc kubenswrapper[4990]: E1203 12:39:52.266168 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:52 crc kubenswrapper[4990]: E1203 12:39:52.266243 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:52 crc kubenswrapper[4990]: I1203 12:39:52.266718 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:39:53 crc kubenswrapper[4990]: E1203 12:39:53.089871 4990 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 12:39:53 crc kubenswrapper[4990]: I1203 12:39:53.140101 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/3.log" Dec 03 12:39:53 crc kubenswrapper[4990]: I1203 12:39:53.142841 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerStarted","Data":"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902"} Dec 03 12:39:53 crc kubenswrapper[4990]: I1203 12:39:53.853577 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6sqk8"] Dec 03 12:39:53 crc kubenswrapper[4990]: I1203 12:39:53.853677 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:53 crc kubenswrapper[4990]: E1203 12:39:53.853770 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:54 crc kubenswrapper[4990]: I1203 12:39:54.146379 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:39:54 crc kubenswrapper[4990]: I1203 12:39:54.264168 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:54 crc kubenswrapper[4990]: I1203 12:39:54.264168 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:54 crc kubenswrapper[4990]: E1203 12:39:54.264288 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:54 crc kubenswrapper[4990]: I1203 12:39:54.264168 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:54 crc kubenswrapper[4990]: E1203 12:39:54.264373 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:54 crc kubenswrapper[4990]: E1203 12:39:54.264446 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:56 crc kubenswrapper[4990]: I1203 12:39:56.264406 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:56 crc kubenswrapper[4990]: I1203 12:39:56.264408 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:56 crc kubenswrapper[4990]: I1203 12:39:56.264486 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:56 crc kubenswrapper[4990]: I1203 12:39:56.264502 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:56 crc kubenswrapper[4990]: E1203 12:39:56.264856 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 12:39:56 crc kubenswrapper[4990]: E1203 12:39:56.265013 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6sqk8" podUID="a061fb3f-ace6-4ab6-a499-b34979186d08" Dec 03 12:39:56 crc kubenswrapper[4990]: E1203 12:39:56.265110 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 12:39:56 crc kubenswrapper[4990]: E1203 12:39:56.265148 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.264437 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.264582 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.264697 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.264783 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.267419 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.267931 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.268276 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.270159 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.270332 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 12:39:58 crc kubenswrapper[4990]: I1203 12:39:58.277884 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.278183 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:02 crc kubenswrapper[4990]: E1203 12:40:02.278348 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:42:04.278317916 +0000 UTC m=+272.420229135 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.279095 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.279255 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.279952 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.288107 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.380721 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.380799 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.384503 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.385189 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.486735 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.495895 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.527601 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:40:02 crc kubenswrapper[4990]: I1203 12:40:02.836282 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podStartSLOduration=128.836262978 podStartE2EDuration="2m8.836262978s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:39:54.174112841 +0000 UTC m=+142.316024070" watchObservedRunningTime="2025-12-03 12:40:02.836262978 +0000 UTC m=+150.978174207" Dec 03 12:40:02 crc kubenswrapper[4990]: W1203 12:40:02.862569 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e4febbfd09c438019d9249fef39c2b8f123353ec22601cede7984cabad15d484 WatchSource:0}: Error finding container e4febbfd09c438019d9249fef39c2b8f123353ec22601cede7984cabad15d484: Status 404 returned error can't find the container with id e4febbfd09c438019d9249fef39c2b8f123353ec22601cede7984cabad15d484 Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.057831 4990 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.098485 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.103568 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.111211 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.111216 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.111528 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.115014 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jfgrs"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.119757 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.126774 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.132946 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqdjl"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.133184 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.134915 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6x7k5"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.135100 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dqrqw"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.135286 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.135706 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.136378 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.136750 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.137239 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.138494 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.138615 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.138636 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.138669 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.138676 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.138785 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.139067 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.139189 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.148343 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-56ffp"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.148885 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.149244 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.149928 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.151785 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-lh8cf"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.153079 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.153108 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.153160 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.153739 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.153202 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.154754 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.154844 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155101 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155152 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155154 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155217 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155486 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155590 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155702 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.155962 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156065 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156182 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156259 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156327 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156353 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156430 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156462 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156518 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156593 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156601 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156682 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.156932 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.157169 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.157197 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.161550 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8f7ss"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.162280 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-89sr2"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.163035 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.168111 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.168385 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.170752 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.171274 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.172884 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.173129 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.173837 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-59lwv"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.174291 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203366 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afded807-a057-4f22-8ff3-88870f086ed1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203768 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-etcd-client\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203802 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-service-ca-bundle\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203835 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnkn8\" (UniqueName: \"kubernetes.io/projected/8cf68834-8149-4d64-8154-d272c920149f-kube-api-access-wnkn8\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203862 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-serving-cert\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203879 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203908 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203933 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203959 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/184ef2da-fe68-4fbc-8a55-f6eb63262764-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.203985 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-config\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.204015 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-policies\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.204037 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.204065 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.204089 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-dir\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.206113 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.209068 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.212493 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.212635 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.211552 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-etcd-serving-ca\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.212800 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-image-import-ca\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.212828 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-auth-proxy-config\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.212754 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.212958 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-machine-approver-tls\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.213020 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7694\" (UniqueName: \"kubernetes.io/projected/afded807-a057-4f22-8ff3-88870f086ed1-kube-api-access-m7694\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.213104 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbw7j\" (UniqueName: \"kubernetes.io/projected/184ef2da-fe68-4fbc-8a55-f6eb63262764-kube-api-access-kbw7j\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.213415 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.213771 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.213921 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184ef2da-fe68-4fbc-8a55-f6eb63262764-config\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.213961 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwtdt\" (UniqueName: \"kubernetes.io/projected/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-kube-api-access-jwtdt\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.214014 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.214879 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.214920 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-config\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.214971 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.214977 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-client-ca\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215248 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/650a3b2c-a031-438d-90e2-2daec51b80e1-serving-cert\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215478 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/650a3b2c-a031-438d-90e2-2daec51b80e1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215539 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-node-pullsecrets\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215577 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215633 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-serving-cert\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215663 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215693 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8cf68834-8149-4d64-8154-d272c920149f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215732 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5splc\" (UniqueName: \"kubernetes.io/projected/80543db8-34e7-48cb-93db-897ae58288d2-kube-api-access-5splc\") pod \"downloads-7954f5f757-lh8cf\" (UID: \"80543db8-34e7-48cb-93db-897ae58288d2\") " pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.215760 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-audit\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.216394 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-config\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.216444 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-serving-cert\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.217136 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8tll\" (UniqueName: \"kubernetes.io/projected/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-kube-api-access-b8tll\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.217486 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-audit-dir\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.217530 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ff45\" (UniqueName: \"kubernetes.io/projected/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-kube-api-access-9ff45\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.217570 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.217835 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-config\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218516 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218643 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218267 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218831 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-encryption-config\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218866 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afded807-a057-4f22-8ff3-88870f086ed1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218927 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cf68834-8149-4d64-8154-d272c920149f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218990 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lqsh\" (UniqueName: \"kubernetes.io/projected/650a3b2c-a031-438d-90e2-2daec51b80e1-kube-api-access-4lqsh\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219019 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cf68834-8149-4d64-8154-d272c920149f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219048 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttwfl\" (UniqueName: \"kubernetes.io/projected/2ccf32a0-8701-4d6f-8c45-d314f98754b3-kube-api-access-ttwfl\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219072 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/184ef2da-fe68-4fbc-8a55-f6eb63262764-images\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219098 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9060dac-4a34-4558-be5e-c16d356839b5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6jkj9\" (UID: \"f9060dac-4a34-4558-be5e-c16d356839b5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219134 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvxp\" (UniqueName: \"kubernetes.io/projected/f9060dac-4a34-4558-be5e-c16d356839b5-kube-api-access-8rvxp\") pod \"cluster-samples-operator-665b6dd947-6jkj9\" (UID: \"f9060dac-4a34-4558-be5e-c16d356839b5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219160 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219189 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqqk2\" (UniqueName: \"kubernetes.io/projected/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-kube-api-access-kqqk2\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.219215 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.218577 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.220219 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.221563 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.220253 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.220320 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.220393 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.220636 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.221653 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.221984 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.222060 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.222368 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.221208 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.222931 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.223206 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.223515 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.223708 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.224018 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.224071 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.224557 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.224755 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.224261 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.225031 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.225233 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.226391 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.240965 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f4ff05eeea03833cc5708486aba5de7181cce6ad85df13e97aeef9222780464d"} Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.241031 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.241371 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e4febbfd09c438019d9249fef39c2b8f123353ec22601cede7984cabad15d484"} Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.241390 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t2dw4"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.241535 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.241875 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.241919 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.243948 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.244850 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.245326 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.246542 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.249060 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.252943 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.255406 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-h7z6x"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.256091 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.256401 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.256564 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.257048 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.257317 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3afc4a1c9fe6b04cadaa1d99c9e471c103adef796e5151254a9a5de633a0550d"} Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.257730 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.258082 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.258616 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.259694 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.260322 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.264646 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.265339 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.266343 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-fgc99"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.267108 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a3221ac7b91f742ba3f81392ce301b5e5b9a1133f5605849d64e03ac4bf3a2ae"} Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.267144 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"61500cc1271fe382e4a31b035da52636d1284025e1509edafab214877bec91cf"} Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.267285 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.268501 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.268628 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.268825 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-twmj9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.269329 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.273811 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.276920 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.277272 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.277496 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.277730 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.277870 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.278596 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.278751 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.278826 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.278874 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.278990 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.279156 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.279220 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.279398 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.281390 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.281809 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.282576 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.288464 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.288963 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.289331 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.292145 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.292482 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.292699 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.293549 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.294633 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.294692 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.294726 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.294867 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.295523 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.297416 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.298169 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.299237 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.299870 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.300861 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.301582 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.302683 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.303511 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gmsz9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.303974 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.304782 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.312539 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.313493 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320521 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320595 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7694\" (UniqueName: \"kubernetes.io/projected/afded807-a057-4f22-8ff3-88870f086ed1-kube-api-access-m7694\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320688 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbw7j\" (UniqueName: \"kubernetes.io/projected/184ef2da-fe68-4fbc-8a55-f6eb63262764-kube-api-access-kbw7j\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320726 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320772 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4956j\" (UniqueName: \"kubernetes.io/projected/cfc3766b-745f-4fe8-b1e5-beb39019ab01-kube-api-access-4956j\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320885 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184ef2da-fe68-4fbc-8a55-f6eb63262764-config\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320918 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwtdt\" (UniqueName: \"kubernetes.io/projected/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-kube-api-access-jwtdt\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320947 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.320984 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.321015 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-trusted-ca-bundle\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.321055 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-config\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.321102 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-client-ca\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.321137 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/650a3b2c-a031-438d-90e2-2daec51b80e1-serving-cert\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.321240 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-config\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.321286 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/650a3b2c-a031-438d-90e2-2daec51b80e1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.322673 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-config\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.330964 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2nnj\" (UniqueName: \"kubernetes.io/projected/d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba-kube-api-access-v2nnj\") pod \"dns-operator-744455d44c-h7z6x\" (UID: \"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba\") " pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331304 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-node-pullsecrets\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331371 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331581 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pl7s\" (UniqueName: \"kubernetes.io/projected/1a2c5a8d-266b-441a-9977-16b138ebcc6f-kube-api-access-7pl7s\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331645 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-serving-cert\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331743 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331789 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8cf68834-8149-4d64-8154-d272c920149f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331864 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-client-ca\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.331995 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5splc\" (UniqueName: \"kubernetes.io/projected/80543db8-34e7-48cb-93db-897ae58288d2-kube-api-access-5splc\") pod \"downloads-7954f5f757-lh8cf\" (UID: \"80543db8-34e7-48cb-93db-897ae58288d2\") " pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332085 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-serving-cert\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332177 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-audit\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332234 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-config\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332324 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-serving-cert\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332369 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-audit-dir\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332425 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ff45\" (UniqueName: \"kubernetes.io/projected/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-kube-api-access-9ff45\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332561 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8tll\" (UniqueName: \"kubernetes.io/projected/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-kube-api-access-b8tll\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332627 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzr2v\" (UniqueName: \"kubernetes.io/projected/25619895-5809-4fe0-baeb-1746885abadd-kube-api-access-kzr2v\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332683 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-config\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332733 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96f89cc4-d412-486d-bc3d-f819b1537477-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332784 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.332827 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25619895-5809-4fe0-baeb-1746885abadd-config\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.322624 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.333635 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-config\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.333730 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f89cc4-d412-486d-bc3d-f819b1537477-config\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.333840 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.333923 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25619895-5809-4fe0-baeb-1746885abadd-trusted-ca\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.334005 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-encryption-config\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.334086 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a2c5a8d-266b-441a-9977-16b138ebcc6f-serving-cert\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.334149 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.335521 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.334151 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afded807-a057-4f22-8ff3-88870f086ed1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336071 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cf68834-8149-4d64-8154-d272c920149f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336120 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96f89cc4-d412-486d-bc3d-f819b1537477-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336168 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lqsh\" (UniqueName: \"kubernetes.io/projected/650a3b2c-a031-438d-90e2-2daec51b80e1-kube-api-access-4lqsh\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336198 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cf68834-8149-4d64-8154-d272c920149f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336278 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/184ef2da-fe68-4fbc-8a55-f6eb63262764-images\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336307 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9060dac-4a34-4558-be5e-c16d356839b5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6jkj9\" (UID: \"f9060dac-4a34-4558-be5e-c16d356839b5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336607 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttwfl\" (UniqueName: \"kubernetes.io/projected/2ccf32a0-8701-4d6f-8c45-d314f98754b3-kube-api-access-ttwfl\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336649 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25619895-5809-4fe0-baeb-1746885abadd-serving-cert\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.337396 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/650a3b2c-a031-438d-90e2-2daec51b80e1-serving-cert\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.338782 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/650a3b2c-a031-438d-90e2-2daec51b80e1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.339193 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.339385 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.339475 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.342118 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-config\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.342304 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-audit\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.338806 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.343533 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-audit-dir\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.343796 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184ef2da-fe68-4fbc-8a55-f6eb63262764-config\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.345562 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-node-pullsecrets\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.346396 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.346535 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cf68834-8149-4d64-8154-d272c920149f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.347084 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8cf68834-8149-4d64-8154-d272c920149f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.347237 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/184ef2da-fe68-4fbc-8a55-f6eb63262764-images\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.336683 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvxp\" (UniqueName: \"kubernetes.io/projected/f9060dac-4a34-4558-be5e-c16d356839b5-kube-api-access-8rvxp\") pod \"cluster-samples-operator-665b6dd947-6jkj9\" (UID: \"f9060dac-4a34-4558-be5e-c16d356839b5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.347641 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.347688 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqqk2\" (UniqueName: \"kubernetes.io/projected/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-kube-api-access-kqqk2\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.347710 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.347747 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-client-ca\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.348323 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-config\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.348550 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-service-ca\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.349495 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.349804 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.350759 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rftgm"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.350990 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.352622 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.352877 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.353439 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-serving-cert\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.355994 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zdzf9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.356514 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9060dac-4a34-4558-be5e-c16d356839b5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6jkj9\" (UID: \"f9060dac-4a34-4558-be5e-c16d356839b5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.358983 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.360422 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.362189 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.362362 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.362987 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afded807-a057-4f22-8ff3-88870f086ed1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.363214 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-etcd-client\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.363716 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba-metrics-tls\") pod \"dns-operator-744455d44c-h7z6x\" (UID: \"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba\") " pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.363796 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-serving-cert\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364332 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-service-ca-bundle\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364624 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnkn8\" (UniqueName: \"kubernetes.io/projected/8cf68834-8149-4d64-8154-d272c920149f-kube-api-access-wnkn8\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364661 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-oauth-config\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364705 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364743 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364776 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364842 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/184ef2da-fe68-4fbc-8a55-f6eb63262764-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.364938 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-config\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365550 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-policies\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365586 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365645 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365682 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-etcd-serving-ca\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365735 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-dir\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365764 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-image-import-ca\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365794 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-auth-proxy-config\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365868 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-machine-approver-tls\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.365940 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-oauth-serving-cert\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.366796 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-service-ca-bundle\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.368030 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-etcd-client\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.368540 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-serving-cert\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.369808 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-etcd-serving-ca\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.369901 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-dir\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.371390 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.371816 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-config\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.371417 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.371836 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-auth-proxy-config\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.372185 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.372375 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-image-import-ca\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.373213 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/184ef2da-fe68-4fbc-8a55-f6eb63262764-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.373479 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.373736 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.374159 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-policies\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.374839 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.376057 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afded807-a057-4f22-8ff3-88870f086ed1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.376118 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bvfbd"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.376371 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.376842 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-machine-approver-tls\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.378425 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afded807-a057-4f22-8ff3-88870f086ed1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.378562 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.378893 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.379009 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.379089 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.379190 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-serving-cert\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.379675 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.381691 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.383863 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-encryption-config\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.387105 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.390550 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8f7ss"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.393268 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dqrqw"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.394700 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6x7k5"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.396004 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-59lwv"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.398721 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.399904 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.400973 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lh8cf"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.402207 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.403579 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.404923 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jfgrs"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.405935 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.406803 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-twmj9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.407997 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.409205 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-h7z6x"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.410530 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-89sr2"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.411663 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.412952 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.413995 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-56ffp"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.415259 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.416509 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.417539 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.418640 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqdjl"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.419730 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.420850 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.422087 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.423197 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.424432 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t2dw4"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.425690 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.428151 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.428408 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bvfbd"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.428444 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z8465"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.429789 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.430708 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-p8mmq"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.431638 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.440840 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gmsz9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.443801 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rftgm"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.446185 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.447062 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.447219 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zdzf9"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.449543 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.450982 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.453331 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z8465"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.454410 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-p8mmq"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.455520 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lsgzp"] Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.456333 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.466170 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467210 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pl7s\" (UniqueName: \"kubernetes.io/projected/1a2c5a8d-266b-441a-9977-16b138ebcc6f-kube-api-access-7pl7s\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467296 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-serving-cert\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467349 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzr2v\" (UniqueName: \"kubernetes.io/projected/25619895-5809-4fe0-baeb-1746885abadd-kube-api-access-kzr2v\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467374 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-config\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467399 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96f89cc4-d412-486d-bc3d-f819b1537477-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467430 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25619895-5809-4fe0-baeb-1746885abadd-config\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467473 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f89cc4-d412-486d-bc3d-f819b1537477-config\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467533 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25619895-5809-4fe0-baeb-1746885abadd-trusted-ca\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467564 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a2c5a8d-266b-441a-9977-16b138ebcc6f-serving-cert\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467600 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96f89cc4-d412-486d-bc3d-f819b1537477-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467644 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25619895-5809-4fe0-baeb-1746885abadd-serving-cert\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467684 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-client-ca\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467797 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-service-ca\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467842 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba-metrics-tls\") pod \"dns-operator-744455d44c-h7z6x\" (UID: \"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba\") " pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467873 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-oauth-config\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467910 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-oauth-serving-cert\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467950 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4956j\" (UniqueName: \"kubernetes.io/projected/cfc3766b-745f-4fe8-b1e5-beb39019ab01-kube-api-access-4956j\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.467982 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-trusted-ca-bundle\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.468010 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-config\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.468397 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2nnj\" (UniqueName: \"kubernetes.io/projected/d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba-kube-api-access-v2nnj\") pod \"dns-operator-744455d44c-h7z6x\" (UID: \"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba\") " pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.468476 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-config\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.469680 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-service-ca\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.469746 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-client-ca\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.470035 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25619895-5809-4fe0-baeb-1746885abadd-trusted-ca\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.470254 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-config\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.470283 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-trusted-ca-bundle\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.470500 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-oauth-serving-cert\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.470922 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25619895-5809-4fe0-baeb-1746885abadd-config\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.471851 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25619895-5809-4fe0-baeb-1746885abadd-serving-cert\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.471972 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a2c5a8d-266b-441a-9977-16b138ebcc6f-serving-cert\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.473021 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba-metrics-tls\") pod \"dns-operator-744455d44c-h7z6x\" (UID: \"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba\") " pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.474701 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-oauth-config\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.475570 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-serving-cert\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.486300 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.506553 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.532961 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.546332 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.551882 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96f89cc4-d412-486d-bc3d-f819b1537477-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.567620 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.570051 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f89cc4-d412-486d-bc3d-f819b1537477-config\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.586406 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.607742 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.626376 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.646269 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.666476 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.687282 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.706258 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.726532 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.746445 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.767103 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.826508 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.846917 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.866683 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.886147 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.905956 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.926520 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.953586 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.966021 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 12:40:03 crc kubenswrapper[4990]: I1203 12:40:03.985676 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.006330 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.026378 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.046529 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.066940 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.086611 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.106883 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.126284 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.146136 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.166204 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.186663 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.206182 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.226843 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.246210 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.266787 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.271634 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"116d183129bc2057b6fc23138154f4e6b8fa05b3043157d14074c4a9e3a65504"} Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.301475 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.305277 4990 request.go:700] Waited for 1.004164017s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&limit=500&resourceVersion=0 Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.308263 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.326952 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.346242 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.366848 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.393680 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.406565 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.426947 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.446686 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.466508 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.487773 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.506083 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.526503 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.566437 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbw7j\" (UniqueName: \"kubernetes.io/projected/184ef2da-fe68-4fbc-8a55-f6eb63262764-kube-api-access-kbw7j\") pod \"machine-api-operator-5694c8668f-jfgrs\" (UID: \"184ef2da-fe68-4fbc-8a55-f6eb63262764\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.579268 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7694\" (UniqueName: \"kubernetes.io/projected/afded807-a057-4f22-8ff3-88870f086ed1-kube-api-access-m7694\") pod \"openshift-apiserver-operator-796bbdcf4f-tk4fc\" (UID: \"afded807-a057-4f22-8ff3-88870f086ed1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.600846 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwtdt\" (UniqueName: \"kubernetes.io/projected/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-kube-api-access-jwtdt\") pod \"controller-manager-879f6c89f-nqdjl\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.608895 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.626404 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.646295 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.649619 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.667178 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.687329 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.721362 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.721978 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ff45\" (UniqueName: \"kubernetes.io/projected/4c2e46b0-facc-41e1-9734-9b05e06fb2f2-kube-api-access-9ff45\") pod \"authentication-operator-69f744f599-6x7k5\" (UID: \"4c2e46b0-facc-41e1-9734-9b05e06fb2f2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.740652 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5splc\" (UniqueName: \"kubernetes.io/projected/80543db8-34e7-48cb-93db-897ae58288d2-kube-api-access-5splc\") pod \"downloads-7954f5f757-lh8cf\" (UID: \"80543db8-34e7-48cb-93db-897ae58288d2\") " pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.752368 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.761550 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8tll\" (UniqueName: \"kubernetes.io/projected/b2518dc1-c310-46c0-9cad-45a9cdf8d39b-kube-api-access-b8tll\") pod \"machine-approver-56656f9798-xs2zs\" (UID: \"b2518dc1-c310-46c0-9cad-45a9cdf8d39b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.776810 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.799060 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cf68834-8149-4d64-8154-d272c920149f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.809090 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lqsh\" (UniqueName: \"kubernetes.io/projected/650a3b2c-a031-438d-90e2-2daec51b80e1-kube-api-access-4lqsh\") pod \"openshift-config-operator-7777fb866f-56ffp\" (UID: \"650a3b2c-a031-438d-90e2-2daec51b80e1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.831642 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttwfl\" (UniqueName: \"kubernetes.io/projected/2ccf32a0-8701-4d6f-8c45-d314f98754b3-kube-api-access-ttwfl\") pod \"oauth-openshift-558db77b4-dqrqw\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.846896 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvxp\" (UniqueName: \"kubernetes.io/projected/f9060dac-4a34-4558-be5e-c16d356839b5-kube-api-access-8rvxp\") pod \"cluster-samples-operator-665b6dd947-6jkj9\" (UID: \"f9060dac-4a34-4558-be5e-c16d356839b5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.863137 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqqk2\" (UniqueName: \"kubernetes.io/projected/8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc-kube-api-access-kqqk2\") pod \"apiserver-76f77b778f-8f7ss\" (UID: \"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc\") " pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.866865 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.886399 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.901909 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.906329 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.915285 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.928213 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.928323 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.937816 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:04 crc kubenswrapper[4990]: W1203 12:40:04.945042 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2518dc1_c310_46c0_9cad_45a9cdf8d39b.slice/crio-5f3cc47b54412d23512f0f5ddce2a4cab52f29ed5b927b42f902b8f6f3094b7e WatchSource:0}: Error finding container 5f3cc47b54412d23512f0f5ddce2a4cab52f29ed5b927b42f902b8f6f3094b7e: Status 404 returned error can't find the container with id 5f3cc47b54412d23512f0f5ddce2a4cab52f29ed5b927b42f902b8f6f3094b7e Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.946258 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.947532 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc"] Dec 03 12:40:04 crc kubenswrapper[4990]: W1203 12:40:04.956128 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafded807_a057_4f22_8ff3_88870f086ed1.slice/crio-ad109b47bf0f7f3844abb1cf3dab9745ea5d25b50ad53053ff27e2abb0cce2d5 WatchSource:0}: Error finding container ad109b47bf0f7f3844abb1cf3dab9745ea5d25b50ad53053ff27e2abb0cce2d5: Status 404 returned error can't find the container with id ad109b47bf0f7f3844abb1cf3dab9745ea5d25b50ad53053ff27e2abb0cce2d5 Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.967247 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.987176 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jfgrs"] Dec 03 12:40:04 crc kubenswrapper[4990]: I1203 12:40:04.989134 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 12:40:05 crc kubenswrapper[4990]: W1203 12:40:04.995594 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod184ef2da_fe68_4fbc_8a55_f6eb63262764.slice/crio-825eb4e3ceb34b847180c5fb34aaee027eda2bd1c276693435533b3c3d955bb8 WatchSource:0}: Error finding container 825eb4e3ceb34b847180c5fb34aaee027eda2bd1c276693435533b3c3d955bb8: Status 404 returned error can't find the container with id 825eb4e3ceb34b847180c5fb34aaee027eda2bd1c276693435533b3c3d955bb8 Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.006642 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.028068 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.046067 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqdjl"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.049347 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.068222 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 12:40:05 crc kubenswrapper[4990]: W1203 12:40:05.073012 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c24e0b4_310e_40c5_8ba6_c2fcd5a6b906.slice/crio-ef29f59bdcbbcb5a0358254e14da115cd36feb4d3c25acdaedf3d642824a78d2 WatchSource:0}: Error finding container ef29f59bdcbbcb5a0358254e14da115cd36feb4d3c25acdaedf3d642824a78d2: Status 404 returned error can't find the container with id ef29f59bdcbbcb5a0358254e14da115cd36feb4d3c25acdaedf3d642824a78d2 Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.077323 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6x7k5"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.103820 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnkn8\" (UniqueName: \"kubernetes.io/projected/8cf68834-8149-4d64-8154-d272c920149f-kube-api-access-wnkn8\") pod \"cluster-image-registry-operator-dc59b4c8b-9cr2s\" (UID: \"8cf68834-8149-4d64-8154-d272c920149f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.105945 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.117300 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.126117 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.135555 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.151985 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.165902 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.180247 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-56ffp"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.190044 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.192730 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.206429 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 12:40:05 crc kubenswrapper[4990]: W1203 12:40:05.213859 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod650a3b2c_a031_438d_90e2_2daec51b80e1.slice/crio-83fdc1151fc2e97aa4964775dc54d9c3b77d3d92a0e73351e1508d6c39f2bf26 WatchSource:0}: Error finding container 83fdc1151fc2e97aa4964775dc54d9c3b77d3d92a0e73351e1508d6c39f2bf26: Status 404 returned error can't find the container with id 83fdc1151fc2e97aa4964775dc54d9c3b77d3d92a0e73351e1508d6c39f2bf26 Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.226549 4990 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.247686 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.257250 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lh8cf"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.269562 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.285025 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" event={"ID":"4c2e46b0-facc-41e1-9734-9b05e06fb2f2","Type":"ContainerStarted","Data":"505593c7e16d88ee0479f9b8a0ff820174c9cb8ead56ca68df537818647ef89e"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.285087 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" event={"ID":"4c2e46b0-facc-41e1-9734-9b05e06fb2f2","Type":"ContainerStarted","Data":"9419803bca684f4d8f4a2a3cfec5721309adb4f1b60f50c414eb1032e365df36"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.287761 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.306593 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.307400 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" event={"ID":"afded807-a057-4f22-8ff3-88870f086ed1","Type":"ContainerStarted","Data":"daec82a5c559270a70558b0aa924b63da531010e7a0bbc966de7a78b1015d639"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.307476 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" event={"ID":"afded807-a057-4f22-8ff3-88870f086ed1","Type":"ContainerStarted","Data":"ad109b47bf0f7f3844abb1cf3dab9745ea5d25b50ad53053ff27e2abb0cce2d5"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.310714 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" event={"ID":"b2518dc1-c310-46c0-9cad-45a9cdf8d39b","Type":"ContainerStarted","Data":"5f3cc47b54412d23512f0f5ddce2a4cab52f29ed5b927b42f902b8f6f3094b7e"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.325333 4990 request.go:700] Waited for 1.893072188s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.328659 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.336842 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8f7ss"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.340040 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" event={"ID":"184ef2da-fe68-4fbc-8a55-f6eb63262764","Type":"ContainerStarted","Data":"a1d5abe84b3b46557e897262a901ed99c53c1ac03bb518030e79bd5551078961"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.340105 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" event={"ID":"184ef2da-fe68-4fbc-8a55-f6eb63262764","Type":"ContainerStarted","Data":"825eb4e3ceb34b847180c5fb34aaee027eda2bd1c276693435533b3c3d955bb8"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.356272 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.365000 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" event={"ID":"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906","Type":"ContainerStarted","Data":"b60c1fc2097278682f43c5ca3d540f8fc5a3c85a967e384607642fa270c602a7"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.365042 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" event={"ID":"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906","Type":"ContainerStarted","Data":"ef29f59bdcbbcb5a0358254e14da115cd36feb4d3c25acdaedf3d642824a78d2"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.366520 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.367333 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.374140 4990 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nqdjl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.374201 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" podUID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.376119 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" event={"ID":"650a3b2c-a031-438d-90e2-2daec51b80e1","Type":"ContainerStarted","Data":"83fdc1151fc2e97aa4964775dc54d9c3b77d3d92a0e73351e1508d6c39f2bf26"} Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.387197 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.406379 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.418527 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dqrqw"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.462642 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzr2v\" (UniqueName: \"kubernetes.io/projected/25619895-5809-4fe0-baeb-1746885abadd-kube-api-access-kzr2v\") pod \"console-operator-58897d9998-89sr2\" (UID: \"25619895-5809-4fe0-baeb-1746885abadd\") " pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.468274 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pl7s\" (UniqueName: \"kubernetes.io/projected/1a2c5a8d-266b-441a-9977-16b138ebcc6f-kube-api-access-7pl7s\") pod \"route-controller-manager-6576b87f9c-25q7v\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.480315 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.489174 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2nnj\" (UniqueName: \"kubernetes.io/projected/d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba-kube-api-access-v2nnj\") pod \"dns-operator-744455d44c-h7z6x\" (UID: \"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba\") " pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.505178 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96f89cc4-d412-486d-bc3d-f819b1537477-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-55tzf\" (UID: \"96f89cc4-d412-486d-bc3d-f819b1537477\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.546059 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.549358 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4956j\" (UniqueName: \"kubernetes.io/projected/cfc3766b-745f-4fe8-b1e5-beb39019ab01-kube-api-access-4956j\") pod \"console-f9d7485db-59lwv\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.561985 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.583112 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.587282 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s"] Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.592151 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605574 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/791d5adb-b1b5-4cf6-a003-f8ad41da334e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605615 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09e7e6a7-b460-4f2d-9c66-28b41381364b-serving-cert\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605636 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-config\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605677 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/791d5adb-b1b5-4cf6-a003-f8ad41da334e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605693 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkcgl\" (UniqueName: \"kubernetes.io/projected/11fe8587-dd60-4149-853e-db89b1cae352-kube-api-access-vkcgl\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605734 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-etcd-client\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605750 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-service-ca\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605773 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-service-ca-bundle\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605787 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-stats-auth\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605804 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605826 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605843 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-metrics-certs\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605882 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-serving-cert\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605919 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-trusted-ca\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605934 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-default-certificate\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605971 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vmg8\" (UniqueName: \"kubernetes.io/projected/09e7e6a7-b460-4f2d-9c66-28b41381364b-kube-api-access-7vmg8\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.605991 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fe8587-dd60-4149-853e-db89b1cae352-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606036 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-encryption-config\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606100 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-client\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606124 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11fe8587-dd60-4149-853e-db89b1cae352-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606168 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-tls\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606201 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606223 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-bound-sa-token\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606242 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqzxd\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-kube-api-access-fqzxd\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606293 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-certificates\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606322 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqmlq\" (UniqueName: \"kubernetes.io/projected/7b8daa8e-d841-4fd6-b34d-74f051da71eb-kube-api-access-mqmlq\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606376 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-audit-policies\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606392 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv522\" (UniqueName: \"kubernetes.io/projected/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-kube-api-access-tv522\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606409 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7b8daa8e-d841-4fd6-b34d-74f051da71eb-audit-dir\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.606468 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-ca\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: E1203 12:40:05.608772 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.108760893 +0000 UTC m=+154.250672122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.646346 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707530 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707760 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-plugins-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707817 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-audit-policies\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707847 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv522\" (UniqueName: \"kubernetes.io/projected/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-kube-api-access-tv522\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707878 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7b8daa8e-d841-4fd6-b34d-74f051da71eb-audit-dir\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707948 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.707981 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708001 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-images\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708021 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb47fbea-eda2-4b31-b64f-424cd42abf4e-serving-cert\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708043 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d608434-0719-4924-9a9d-7f03de7b6cde-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708062 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708097 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/791d5adb-b1b5-4cf6-a003-f8ad41da334e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708119 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbdfce6d-1752-4940-a0e5-ee5bfa89d733-cert\") pod \"ingress-canary-p8mmq\" (UID: \"fbdfce6d-1752-4940-a0e5-ee5bfa89d733\") " pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708138 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzvgn\" (UniqueName: \"kubernetes.io/projected/7c2e0256-7681-49d7-afbc-145335bf5b17-kube-api-access-lzvgn\") pod \"multus-admission-controller-857f4d67dd-zdzf9\" (UID: \"7c2e0256-7681-49d7-afbc-145335bf5b17\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708170 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb47fbea-eda2-4b31-b64f-424cd42abf4e-config\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708215 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708236 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8387a72-f04e-4004-82c8-5634a54600af-config-volume\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708258 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvqpq\" (UniqueName: \"kubernetes.io/projected/a36bdb31-4352-4ef2-9d56-85cd07b89e21-kube-api-access-fvqpq\") pod \"control-plane-machine-set-operator-78cbb6b69f-jsjk4\" (UID: \"a36bdb31-4352-4ef2-9d56-85cd07b89e21\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708280 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-etcd-client\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708302 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5t4t\" (UniqueName: \"kubernetes.io/projected/0fee873c-8deb-4f94-aa19-632e04acd7de-kube-api-access-s5t4t\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708347 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-stats-auth\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708366 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708403 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-service-ca\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708424 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-metrics-certs\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708445 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fee873c-8deb-4f94-aa19-632e04acd7de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708481 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-metrics-tls\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708500 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-config\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708532 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fee873c-8deb-4f94-aa19-632e04acd7de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708553 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fc9eb76-3dd5-49ed-be37-a5923f4de448-profile-collector-cert\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708573 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq4f9\" (UniqueName: \"kubernetes.io/projected/2870df3c-124c-4a2d-8f88-4fec5f65740c-kube-api-access-vq4f9\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708668 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vmg8\" (UniqueName: \"kubernetes.io/projected/09e7e6a7-b460-4f2d-9c66-28b41381364b-kube-api-access-7vmg8\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708689 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfwfz\" (UniqueName: \"kubernetes.io/projected/bc1ccf16-c231-485b-9e14-8378f725292d-kube-api-access-nfwfz\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708715 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-client\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708759 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-bound-sa-token\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708777 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-socket-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708809 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d608434-0719-4924-9a9d-7f03de7b6cde-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708830 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/431416aa-9fd4-4c6a-8668-9e86d8c256aa-webhook-cert\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708873 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/383ca895-74a4-4dd3-a431-e0caadfa246c-node-bootstrap-token\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708893 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708935 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rfrb\" (UniqueName: \"kubernetes.io/projected/383ca895-74a4-4dd3-a431-e0caadfa246c-kube-api-access-8rfrb\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708954 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj7q8\" (UniqueName: \"kubernetes.io/projected/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-kube-api-access-pj7q8\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708974 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-registration-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.708993 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-signing-key\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709016 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqmlq\" (UniqueName: \"kubernetes.io/projected/7b8daa8e-d841-4fd6-b34d-74f051da71eb-kube-api-access-mqmlq\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709036 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-mountpoint-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709055 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-trusted-ca\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709073 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/383ca895-74a4-4dd3-a431-e0caadfa246c-certs\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709106 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bdc9a406-d80a-4122-b556-5e0c1a12177b-proxy-tls\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709125 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8387a72-f04e-4004-82c8-5634a54600af-metrics-tls\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709147 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-signing-cabundle\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709165 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkmnh\" (UniqueName: \"kubernetes.io/projected/5fc9eb76-3dd5-49ed-be37-a5923f4de448-kube-api-access-gkmnh\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709185 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a36bdb31-4352-4ef2-9d56-85cd07b89e21-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jsjk4\" (UID: \"a36bdb31-4352-4ef2-9d56-85cd07b89e21\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709205 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-csi-data-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709234 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-ca\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709267 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/791d5adb-b1b5-4cf6-a003-f8ad41da334e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709288 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09e7e6a7-b460-4f2d-9c66-28b41381364b-serving-cert\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709306 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-config\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709328 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxlp8\" (UniqueName: \"kubernetes.io/projected/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-kube-api-access-sxlp8\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709350 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkcgl\" (UniqueName: \"kubernetes.io/projected/11fe8587-dd60-4149-853e-db89b1cae352-kube-api-access-vkcgl\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709371 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq4db\" (UniqueName: \"kubernetes.io/projected/bdc9a406-d80a-4122-b556-5e0c1a12177b-kube-api-access-zq4db\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709392 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrz7l\" (UniqueName: \"kubernetes.io/projected/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-kube-api-access-jrz7l\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709411 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709444 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c2e0256-7681-49d7-afbc-145335bf5b17-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zdzf9\" (UID: \"7c2e0256-7681-49d7-afbc-145335bf5b17\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709604 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-proxy-tls\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709642 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dcf4e387-1c02-4274-a26c-c7ae9085f24f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2d2c8\" (UID: \"dcf4e387-1c02-4274-a26c-c7ae9085f24f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709690 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-service-ca-bundle\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709711 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/826bd110-77e7-484c-bd48-d95fa1b4a0cd-secret-volume\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709732 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmq6k\" (UniqueName: \"kubernetes.io/projected/826bd110-77e7-484c-bd48-d95fa1b4a0cd-kube-api-access-xmq6k\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709781 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-srv-cert\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709813 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-serving-cert\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709831 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/431416aa-9fd4-4c6a-8668-9e86d8c256aa-tmpfs\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709851 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmclr\" (UniqueName: \"kubernetes.io/projected/e8387a72-f04e-4004-82c8-5634a54600af-kube-api-access-kmclr\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709874 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs7z6\" (UniqueName: \"kubernetes.io/projected/dcf4e387-1c02-4274-a26c-c7ae9085f24f-kube-api-access-zs7z6\") pod \"package-server-manager-789f6589d5-2d2c8\" (UID: \"dcf4e387-1c02-4274-a26c-c7ae9085f24f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709904 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fc9eb76-3dd5-49ed-be37-a5923f4de448-srv-cert\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709951 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-trusted-ca\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709970 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/431416aa-9fd4-4c6a-8668-9e86d8c256aa-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.709993 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-default-certificate\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710016 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fe8587-dd60-4149-853e-db89b1cae352-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710036 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-encryption-config\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710055 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d608434-0719-4924-9a9d-7f03de7b6cde-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710104 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11fe8587-dd60-4149-853e-db89b1cae352-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710140 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bdc9a406-d80a-4122-b556-5e0c1a12177b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710161 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-tls\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710181 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/826bd110-77e7-484c-bd48-d95fa1b4a0cd-config-volume\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710205 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqzxd\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-kube-api-access-fqzxd\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710225 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710248 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz75r\" (UniqueName: \"kubernetes.io/projected/fbdfce6d-1752-4940-a0e5-ee5bfa89d733-kube-api-access-zz75r\") pod \"ingress-canary-p8mmq\" (UID: \"fbdfce6d-1752-4940-a0e5-ee5bfa89d733\") " pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710320 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qchm7\" (UniqueName: \"kubernetes.io/projected/cb47fbea-eda2-4b31-b64f-424cd42abf4e-kube-api-access-qchm7\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710342 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710362 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5wg7\" (UniqueName: \"kubernetes.io/projected/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-kube-api-access-h5wg7\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710425 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-certificates\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710446 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mn75\" (UniqueName: \"kubernetes.io/projected/431416aa-9fd4-4c6a-8668-9e86d8c256aa-kube-api-access-4mn75\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.710593 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-478nh\" (UniqueName: \"kubernetes.io/projected/a1f76c88-6b38-4c91-a6db-aa9e35e4672a-kube-api-access-478nh\") pod \"migrator-59844c95c7-fqfvb\" (UID: \"a1f76c88-6b38-4c91-a6db-aa9e35e4672a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" Dec 03 12:40:05 crc kubenswrapper[4990]: E1203 12:40:05.711310 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.211290703 +0000 UTC m=+154.353201932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.712375 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-audit-policies\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.712517 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7b8daa8e-d841-4fd6-b34d-74f051da71eb-audit-dir\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.713919 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/791d5adb-b1b5-4cf6-a003-f8ad41da334e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.716161 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-config\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.718865 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.719595 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-service-ca\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.722022 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-service-ca-bundle\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.723034 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-stats-auth\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.734208 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-trusted-ca\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.737356 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11fe8587-dd60-4149-853e-db89b1cae352-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.737679 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-client\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.738139 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-serving-cert\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.738661 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-metrics-certs\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.739031 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09e7e6a7-b460-4f2d-9c66-28b41381364b-serving-cert\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.739541 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-tls\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.739938 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-etcd-client\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.753519 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fe8587-dd60-4149-853e-db89b1cae352-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.754318 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09e7e6a7-b460-4f2d-9c66-28b41381364b-etcd-ca\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.754973 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-default-certificate\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.756402 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-certificates\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.760615 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7b8daa8e-d841-4fd6-b34d-74f051da71eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.787643 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7b8daa8e-d841-4fd6-b34d-74f051da71eb-encryption-config\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.787962 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv522\" (UniqueName: \"kubernetes.io/projected/21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c-kube-api-access-tv522\") pod \"router-default-5444994796-fgc99\" (UID: \"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c\") " pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.788106 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/791d5adb-b1b5-4cf6-a003-f8ad41da334e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.805868 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vmg8\" (UniqueName: \"kubernetes.io/projected/09e7e6a7-b460-4f2d-9c66-28b41381364b-kube-api-access-7vmg8\") pod \"etcd-operator-b45778765-t2dw4\" (UID: \"09e7e6a7-b460-4f2d-9c66-28b41381364b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811312 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811358 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8387a72-f04e-4004-82c8-5634a54600af-config-volume\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811383 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvqpq\" (UniqueName: \"kubernetes.io/projected/a36bdb31-4352-4ef2-9d56-85cd07b89e21-kube-api-access-fvqpq\") pod \"control-plane-machine-set-operator-78cbb6b69f-jsjk4\" (UID: \"a36bdb31-4352-4ef2-9d56-85cd07b89e21\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811407 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5t4t\" (UniqueName: \"kubernetes.io/projected/0fee873c-8deb-4f94-aa19-632e04acd7de-kube-api-access-s5t4t\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811434 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fee873c-8deb-4f94-aa19-632e04acd7de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811477 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-metrics-tls\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811497 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-config\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811522 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811541 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fee873c-8deb-4f94-aa19-632e04acd7de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811559 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fc9eb76-3dd5-49ed-be37-a5923f4de448-profile-collector-cert\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811589 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq4f9\" (UniqueName: \"kubernetes.io/projected/2870df3c-124c-4a2d-8f88-4fec5f65740c-kube-api-access-vq4f9\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811626 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfwfz\" (UniqueName: \"kubernetes.io/projected/bc1ccf16-c231-485b-9e14-8378f725292d-kube-api-access-nfwfz\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811662 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-socket-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811686 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d608434-0719-4924-9a9d-7f03de7b6cde-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811712 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/431416aa-9fd4-4c6a-8668-9e86d8c256aa-webhook-cert\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811733 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/383ca895-74a4-4dd3-a431-e0caadfa246c-node-bootstrap-token\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811753 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811774 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rfrb\" (UniqueName: \"kubernetes.io/projected/383ca895-74a4-4dd3-a431-e0caadfa246c-kube-api-access-8rfrb\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811798 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj7q8\" (UniqueName: \"kubernetes.io/projected/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-kube-api-access-pj7q8\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811819 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-registration-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811844 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-signing-key\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811881 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-mountpoint-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811903 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-trusted-ca\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811922 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/383ca895-74a4-4dd3-a431-e0caadfa246c-certs\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811945 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bdc9a406-d80a-4122-b556-5e0c1a12177b-proxy-tls\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811970 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8387a72-f04e-4004-82c8-5634a54600af-metrics-tls\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.811991 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-signing-cabundle\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.812012 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkmnh\" (UniqueName: \"kubernetes.io/projected/5fc9eb76-3dd5-49ed-be37-a5923f4de448-kube-api-access-gkmnh\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.812032 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a36bdb31-4352-4ef2-9d56-85cd07b89e21-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jsjk4\" (UID: \"a36bdb31-4352-4ef2-9d56-85cd07b89e21\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.812056 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-csi-data-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.812090 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxlp8\" (UniqueName: \"kubernetes.io/projected/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-kube-api-access-sxlp8\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.812116 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq4db\" (UniqueName: \"kubernetes.io/projected/bdc9a406-d80a-4122-b556-5e0c1a12177b-kube-api-access-zq4db\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.812141 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrz7l\" (UniqueName: \"kubernetes.io/projected/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-kube-api-access-jrz7l\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.816686 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-trusted-ca\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.817187 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.817571 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c2e0256-7681-49d7-afbc-145335bf5b17-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zdzf9\" (UID: \"7c2e0256-7681-49d7-afbc-145335bf5b17\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.818100 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8387a72-f04e-4004-82c8-5634a54600af-config-volume\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819496 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-registration-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819774 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-proxy-tls\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819811 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dcf4e387-1c02-4274-a26c-c7ae9085f24f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2d2c8\" (UID: \"dcf4e387-1c02-4274-a26c-c7ae9085f24f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819844 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/826bd110-77e7-484c-bd48-d95fa1b4a0cd-secret-volume\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819869 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmq6k\" (UniqueName: \"kubernetes.io/projected/826bd110-77e7-484c-bd48-d95fa1b4a0cd-kube-api-access-xmq6k\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819907 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-srv-cert\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819933 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/431416aa-9fd4-4c6a-8668-9e86d8c256aa-tmpfs\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819953 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmclr\" (UniqueName: \"kubernetes.io/projected/e8387a72-f04e-4004-82c8-5634a54600af-kube-api-access-kmclr\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.819980 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs7z6\" (UniqueName: \"kubernetes.io/projected/dcf4e387-1c02-4274-a26c-c7ae9085f24f-kube-api-access-zs7z6\") pod \"package-server-manager-789f6589d5-2d2c8\" (UID: \"dcf4e387-1c02-4274-a26c-c7ae9085f24f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820004 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fc9eb76-3dd5-49ed-be37-a5923f4de448-srv-cert\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820039 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/431416aa-9fd4-4c6a-8668-9e86d8c256aa-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820072 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d608434-0719-4924-9a9d-7f03de7b6cde-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820105 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bdc9a406-d80a-4122-b556-5e0c1a12177b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820129 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/826bd110-77e7-484c-bd48-d95fa1b4a0cd-config-volume\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820162 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz75r\" (UniqueName: \"kubernetes.io/projected/fbdfce6d-1752-4940-a0e5-ee5bfa89d733-kube-api-access-zz75r\") pod \"ingress-canary-p8mmq\" (UID: \"fbdfce6d-1752-4940-a0e5-ee5bfa89d733\") " pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820192 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qchm7\" (UniqueName: \"kubernetes.io/projected/cb47fbea-eda2-4b31-b64f-424cd42abf4e-kube-api-access-qchm7\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820212 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820233 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5wg7\" (UniqueName: \"kubernetes.io/projected/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-kube-api-access-h5wg7\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820265 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mn75\" (UniqueName: \"kubernetes.io/projected/431416aa-9fd4-4c6a-8668-9e86d8c256aa-kube-api-access-4mn75\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820290 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-478nh\" (UniqueName: \"kubernetes.io/projected/a1f76c88-6b38-4c91-a6db-aa9e35e4672a-kube-api-access-478nh\") pod \"migrator-59844c95c7-fqfvb\" (UID: \"a1f76c88-6b38-4c91-a6db-aa9e35e4672a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820317 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-plugins-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820353 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820381 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820403 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-images\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820424 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb47fbea-eda2-4b31-b64f-424cd42abf4e-serving-cert\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820497 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d608434-0719-4924-9a9d-7f03de7b6cde-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820526 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820554 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbdfce6d-1752-4940-a0e5-ee5bfa89d733-cert\") pod \"ingress-canary-p8mmq\" (UID: \"fbdfce6d-1752-4940-a0e5-ee5bfa89d733\") " pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820575 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzvgn\" (UniqueName: \"kubernetes.io/projected/7c2e0256-7681-49d7-afbc-145335bf5b17-kube-api-access-lzvgn\") pod \"multus-admission-controller-857f4d67dd-zdzf9\" (UID: \"7c2e0256-7681-49d7-afbc-145335bf5b17\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.820598 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb47fbea-eda2-4b31-b64f-424cd42abf4e-config\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.821493 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-mountpoint-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: E1203 12:40:05.821863 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.321845319 +0000 UTC m=+154.463756628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.822225 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fee873c-8deb-4f94-aa19-632e04acd7de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.823103 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-config\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.829233 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d608434-0719-4924-9a9d-7f03de7b6cde-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.830417 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/826bd110-77e7-484c-bd48-d95fa1b4a0cd-config-volume\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.834574 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-socket-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.836083 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d608434-0719-4924-9a9d-7f03de7b6cde-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.836472 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb47fbea-eda2-4b31-b64f-424cd42abf4e-config\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.836691 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-metrics-tls\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.837768 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/431416aa-9fd4-4c6a-8668-9e86d8c256aa-tmpfs\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.838712 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/383ca895-74a4-4dd3-a431-e0caadfa246c-certs\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.844285 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8387a72-f04e-4004-82c8-5634a54600af-metrics-tls\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.845153 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-signing-cabundle\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.845801 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bdc9a406-d80a-4122-b556-5e0c1a12177b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.846013 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.850257 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fc9eb76-3dd5-49ed-be37-a5923f4de448-profile-collector-cert\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.850497 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-csi-data-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.850435 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/bc1ccf16-c231-485b-9e14-8378f725292d-plugins-dir\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.852038 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-srv-cert\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.852371 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-images\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.854702 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-bound-sa-token\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.854808 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c2e0256-7681-49d7-afbc-145335bf5b17-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zdzf9\" (UID: \"7c2e0256-7681-49d7-afbc-145335bf5b17\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.857943 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkcgl\" (UniqueName: \"kubernetes.io/projected/11fe8587-dd60-4149-853e-db89b1cae352-kube-api-access-vkcgl\") pod \"openshift-controller-manager-operator-756b6f6bc6-7vjl9\" (UID: \"11fe8587-dd60-4149-853e-db89b1cae352\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.858612 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.861152 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bdc9a406-d80a-4122-b556-5e0c1a12177b-proxy-tls\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.866882 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.867004 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dcf4e387-1c02-4274-a26c-c7ae9085f24f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2d2c8\" (UID: \"dcf4e387-1c02-4274-a26c-c7ae9085f24f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.869517 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-proxy-tls\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.869572 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.874185 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.874421 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.883295 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/431416aa-9fd4-4c6a-8668-9e86d8c256aa-webhook-cert\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.883805 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb47fbea-eda2-4b31-b64f-424cd42abf4e-serving-cert\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.883866 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/826bd110-77e7-484c-bd48-d95fa1b4a0cd-secret-volume\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.884321 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-signing-key\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.884376 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a36bdb31-4352-4ef2-9d56-85cd07b89e21-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jsjk4\" (UID: \"a36bdb31-4352-4ef2-9d56-85cd07b89e21\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.885122 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/383ca895-74a4-4dd3-a431-e0caadfa246c-node-bootstrap-token\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.885217 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fc9eb76-3dd5-49ed-be37-a5923f4de448-srv-cert\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.886775 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqmlq\" (UniqueName: \"kubernetes.io/projected/7b8daa8e-d841-4fd6-b34d-74f051da71eb-kube-api-access-mqmlq\") pod \"apiserver-7bbb656c7d-qxchk\" (UID: \"7b8daa8e-d841-4fd6-b34d-74f051da71eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.888035 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbdfce6d-1752-4940-a0e5-ee5bfa89d733-cert\") pod \"ingress-canary-p8mmq\" (UID: \"fbdfce6d-1752-4940-a0e5-ee5bfa89d733\") " pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.894328 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/431416aa-9fd4-4c6a-8668-9e86d8c256aa-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.900813 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqzxd\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-kube-api-access-fqzxd\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.926596 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.927889 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:05 crc kubenswrapper[4990]: E1203 12:40:05.928568 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.428546215 +0000 UTC m=+154.570457444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.940047 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fee873c-8deb-4f94-aa19-632e04acd7de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.940434 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.956228 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.968780 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0e245e7d-f3ac-46c3-9186-e7a8fffc02ad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lthk\" (UID: \"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.989266 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5t4t\" (UniqueName: \"kubernetes.io/projected/0fee873c-8deb-4f94-aa19-632e04acd7de-kube-api-access-s5t4t\") pod \"kube-storage-version-migrator-operator-b67b599dd-5f4j8\" (UID: \"0fee873c-8deb-4f94-aa19-632e04acd7de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:05 crc kubenswrapper[4990]: I1203 12:40:05.992389 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj7q8\" (UniqueName: \"kubernetes.io/projected/ef0fa8d8-e755-474d-bdf2-2f742d4a87ff-kube-api-access-pj7q8\") pod \"machine-config-operator-74547568cd-kdq68\" (UID: \"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.002367 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.007065 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvqpq\" (UniqueName: \"kubernetes.io/projected/a36bdb31-4352-4ef2-9d56-85cd07b89e21-kube-api-access-fvqpq\") pod \"control-plane-machine-set-operator-78cbb6b69f-jsjk4\" (UID: \"a36bdb31-4352-4ef2-9d56-85cd07b89e21\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.028008 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-478nh\" (UniqueName: \"kubernetes.io/projected/a1f76c88-6b38-4c91-a6db-aa9e35e4672a-kube-api-access-478nh\") pod \"migrator-59844c95c7-fqfvb\" (UID: \"a1f76c88-6b38-4c91-a6db-aa9e35e4672a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.028417 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfwfz\" (UniqueName: \"kubernetes.io/projected/bc1ccf16-c231-485b-9e14-8378f725292d-kube-api-access-nfwfz\") pod \"csi-hostpathplugin-z8465\" (UID: \"bc1ccf16-c231-485b-9e14-8378f725292d\") " pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.030054 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.030537 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.53052439 +0000 UTC m=+154.672435619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.046199 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz75r\" (UniqueName: \"kubernetes.io/projected/fbdfce6d-1752-4940-a0e5-ee5bfa89d733-kube-api-access-zz75r\") pod \"ingress-canary-p8mmq\" (UID: \"fbdfce6d-1752-4940-a0e5-ee5bfa89d733\") " pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.065200 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq4f9\" (UniqueName: \"kubernetes.io/projected/2870df3c-124c-4a2d-8f88-4fec5f65740c-kube-api-access-vq4f9\") pod \"marketplace-operator-79b997595-gmsz9\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.065775 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.114733 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.115547 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.117261 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzvgn\" (UniqueName: \"kubernetes.io/projected/7c2e0256-7681-49d7-afbc-145335bf5b17-kube-api-access-lzvgn\") pod \"multus-admission-controller-857f4d67dd-zdzf9\" (UID: \"7c2e0256-7681-49d7-afbc-145335bf5b17\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.122693 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.131988 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.132128 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.132573 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.632553245 +0000 UTC m=+154.774464474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.135222 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d608434-0719-4924-9a9d-7f03de7b6cde-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vb5gm\" (UID: \"2d608434-0719-4924-9a9d-7f03de7b6cde\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.135910 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmq6k\" (UniqueName: \"kubernetes.io/projected/826bd110-77e7-484c-bd48-d95fa1b4a0cd-kube-api-access-xmq6k\") pod \"collect-profiles-29412750-l877l\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.139909 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.163103 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmclr\" (UniqueName: \"kubernetes.io/projected/e8387a72-f04e-4004-82c8-5634a54600af-kube-api-access-kmclr\") pod \"dns-default-bvfbd\" (UID: \"e8387a72-f04e-4004-82c8-5634a54600af\") " pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.172828 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs7z6\" (UniqueName: \"kubernetes.io/projected/dcf4e387-1c02-4274-a26c-c7ae9085f24f-kube-api-access-zs7z6\") pod \"package-server-manager-789f6589d5-2d2c8\" (UID: \"dcf4e387-1c02-4274-a26c-c7ae9085f24f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.173414 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z8465" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.184118 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-p8mmq" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.206902 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkmnh\" (UniqueName: \"kubernetes.io/projected/5fc9eb76-3dd5-49ed-be37-a5923f4de448-kube-api-access-gkmnh\") pod \"catalog-operator-68c6474976-8ctlj\" (UID: \"5fc9eb76-3dd5-49ed-be37-a5923f4de448\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.234964 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.237170 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.737150048 +0000 UTC m=+154.879061277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.250365 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq4db\" (UniqueName: \"kubernetes.io/projected/bdc9a406-d80a-4122-b556-5e0c1a12177b-kube-api-access-zq4db\") pod \"machine-config-controller-84d6567774-6c8ms\" (UID: \"bdc9a406-d80a-4122-b556-5e0c1a12177b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.263623 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxlp8\" (UniqueName: \"kubernetes.io/projected/214b5d9c-7933-4ec3-b2ae-60ed6d752cfe-kube-api-access-sxlp8\") pod \"service-ca-9c57cc56f-rftgm\" (UID: \"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe\") " pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.278309 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.284811 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qchm7\" (UniqueName: \"kubernetes.io/projected/cb47fbea-eda2-4b31-b64f-424cd42abf4e-kube-api-access-qchm7\") pod \"service-ca-operator-777779d784-tp6lf\" (UID: \"cb47fbea-eda2-4b31-b64f-424cd42abf4e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.285759 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.293919 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rfrb\" (UniqueName: \"kubernetes.io/projected/383ca895-74a4-4dd3-a431-e0caadfa246c-kube-api-access-8rfrb\") pod \"machine-config-server-lsgzp\" (UID: \"383ca895-74a4-4dd3-a431-e0caadfa246c\") " pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.315005 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrz7l\" (UniqueName: \"kubernetes.io/projected/acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f-kube-api-access-jrz7l\") pod \"olm-operator-6b444d44fb-p2clc\" (UID: \"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.322595 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.339803 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.348834 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.357211 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.358609 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.858555843 +0000 UTC m=+155.000467072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.364294 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.364913 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5wg7\" (UniqueName: \"kubernetes.io/projected/a1fb2e05-1e50-4d85-baa7-fb6dc003b53c-kube-api-access-h5wg7\") pod \"ingress-operator-5b745b69d9-kqf6f\" (UID: \"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.365133 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.365286 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.865272386 +0000 UTC m=+155.007183615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.368874 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mn75\" (UniqueName: \"kubernetes.io/projected/431416aa-9fd4-4c6a-8668-9e86d8c256aa-kube-api-access-4mn75\") pod \"packageserver-d55dfcdfc-pk52k\" (UID: \"431416aa-9fd4-4c6a-8668-9e86d8c256aa\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.399278 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.405221 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.422334 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v"] Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.423370 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.447070 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.458739 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lh8cf" event={"ID":"80543db8-34e7-48cb-93db-897ae58288d2","Type":"ContainerStarted","Data":"d639693cc9fc362114eaaf524724196a18b9e0b657416d2ed43f5cb1f4de098e"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.458783 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lh8cf" event={"ID":"80543db8-34e7-48cb-93db-897ae58288d2","Type":"ContainerStarted","Data":"bdc110ec716c057c2d00529dae093646152e5bff97690a5d8753dfb44f2b1d10"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.459209 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.480619 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.480668 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.482005 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.482632 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:06.982609806 +0000 UTC m=+155.124521035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.488281 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lsgzp" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.531531 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" event={"ID":"8cf68834-8149-4d64-8154-d272c920149f","Type":"ContainerStarted","Data":"6a01b0aba302ca67da3fba5c10b785f739f123e5b3b1490a679d2535045bea0d"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.531616 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" event={"ID":"8cf68834-8149-4d64-8154-d272c920149f","Type":"ContainerStarted","Data":"db1069b7cbf88ef9fc5ed47f29b32340141ab1312122eb0e59a8b8284cb6b1e4"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.535931 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-59lwv"] Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.582319 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-89sr2"] Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.583111 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.583532 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.083517824 +0000 UTC m=+155.225429053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.619510 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.627746 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-fgc99" event={"ID":"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c","Type":"ContainerStarted","Data":"31468a93ca061aa065d970953408f009f8532cbbba0599c74875afa0125961ba"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.634995 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.673207 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" event={"ID":"184ef2da-fe68-4fbc-8a55-f6eb63262764","Type":"ContainerStarted","Data":"ff29b4226ac1461b047a9cec029abe4fac622be2453f8055f14237ab8ff13c5c"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.688056 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.689366 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.189346868 +0000 UTC m=+155.331258097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.727098 4990 generic.go:334] "Generic (PLEG): container finished" podID="8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc" containerID="10429a4bb7e9179cea1bcab9d2f5505b311c1d29465e2b8dcf1f068427506ccc" exitCode=0 Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.727186 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" event={"ID":"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc","Type":"ContainerDied","Data":"10429a4bb7e9179cea1bcab9d2f5505b311c1d29465e2b8dcf1f068427506ccc"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.727212 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" event={"ID":"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc","Type":"ContainerStarted","Data":"6db9795c4abe6df91df91443a16735e546823c3b6a8180255db0cb9121ccdad0"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.749837 4990 generic.go:334] "Generic (PLEG): container finished" podID="650a3b2c-a031-438d-90e2-2daec51b80e1" containerID="67fb099d6660f8ff8a23f46b7740ba7a1d04b9988bca9f324e495240607a5dde" exitCode=0 Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.752596 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" event={"ID":"650a3b2c-a031-438d-90e2-2daec51b80e1","Type":"ContainerDied","Data":"67fb099d6660f8ff8a23f46b7740ba7a1d04b9988bca9f324e495240607a5dde"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.792897 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6x7k5" podStartSLOduration=132.792880603 podStartE2EDuration="2m12.792880603s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:06.69988226 +0000 UTC m=+154.841793479" watchObservedRunningTime="2025-12-03 12:40:06.792880603 +0000 UTC m=+154.934791832" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.793572 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.793909 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.293897509 +0000 UTC m=+155.435808728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.794556 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-h7z6x"] Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.803148 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf"] Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.811236 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" event={"ID":"f9060dac-4a34-4558-be5e-c16d356839b5","Type":"ContainerStarted","Data":"58ed93d9bb0ce86e099d4a960b585cbd04010adc8913522e8eb327917b30d93b"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.811278 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" event={"ID":"f9060dac-4a34-4558-be5e-c16d356839b5","Type":"ContainerStarted","Data":"5bdfeb919d4180002b15ad6057aa7172edc6a2b8b2a79f880e0b36e89e45303f"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.826288 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" event={"ID":"2ccf32a0-8701-4d6f-8c45-d314f98754b3","Type":"ContainerStarted","Data":"cc935d040ebc2ca4be0665585ec1a53eb10a7f990431c939655aca43286b536b"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.826348 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" event={"ID":"2ccf32a0-8701-4d6f-8c45-d314f98754b3","Type":"ContainerStarted","Data":"c7f142e6167dfaedad8b9a618b442c8e54615e64d81e617f286b1362313d2caa"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.826964 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.841826 4990 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dqrqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.841875 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 03 12:40:06 crc kubenswrapper[4990]: W1203 12:40:06.850663 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96f89cc4_d412_486d_bc3d_f819b1537477.slice/crio-a37e43d1af5f4c5d787b1b56569db6044567209fab46bd1a8fc6f44987ef0f48 WatchSource:0}: Error finding container a37e43d1af5f4c5d787b1b56569db6044567209fab46bd1a8fc6f44987ef0f48: Status 404 returned error can't find the container with id a37e43d1af5f4c5d787b1b56569db6044567209fab46bd1a8fc6f44987ef0f48 Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.857194 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" event={"ID":"b2518dc1-c310-46c0-9cad-45a9cdf8d39b","Type":"ContainerStarted","Data":"f00fd9d2ad94ca74770361d50b85d06266df2295631510fc5eea55e5121d14cc"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.857867 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" event={"ID":"b2518dc1-c310-46c0-9cad-45a9cdf8d39b","Type":"ContainerStarted","Data":"4e8c38bfdc8d2c2db18d698b134cdd029e35ba6eb4df96d5ff7f2403f387672e"} Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.889641 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" podStartSLOduration=132.889620743 podStartE2EDuration="2m12.889620743s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:06.888795842 +0000 UTC m=+155.030707071" watchObservedRunningTime="2025-12-03 12:40:06.889620743 +0000 UTC m=+155.031531992" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.894929 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:06 crc kubenswrapper[4990]: E1203 12:40:06.896494 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.396471179 +0000 UTC m=+155.538382468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.935442 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:40:06 crc kubenswrapper[4990]: I1203 12:40:06.935491 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t2dw4"] Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.012304 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.013134 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.513120102 +0000 UTC m=+155.655031401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.114096 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.114709 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.614689276 +0000 UTC m=+155.756600505 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.215834 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.216151 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.716140997 +0000 UTC m=+155.858052226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.318790 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.319077 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.819059277 +0000 UTC m=+155.960970506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.477985 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.478647 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:07.978630604 +0000 UTC m=+156.120541823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.625650 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.626422 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.126399828 +0000 UTC m=+156.268311057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.675886 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9"] Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.680692 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z8465"] Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.704123 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-lh8cf" podStartSLOduration=133.704108628 podStartE2EDuration="2m13.704108628s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:07.702716853 +0000 UTC m=+155.844628082" watchObservedRunningTime="2025-12-03 12:40:07.704108628 +0000 UTC m=+155.846019857" Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.725391 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk"] Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.757645 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.758067 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.258054296 +0000 UTC m=+156.399965525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.860391 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.861252 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.361227022 +0000 UTC m=+156.503138251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.861554 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.862015 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.362004763 +0000 UTC m=+156.503915992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.955707 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" event={"ID":"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad","Type":"ContainerStarted","Data":"9415b96c198626270da937adbf9b2ba6e11b948016207ec3d3ae8d34cfc17a9d"} Dec 03 12:40:07 crc kubenswrapper[4990]: I1203 12:40:07.966528 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:07 crc kubenswrapper[4990]: E1203 12:40:07.966960 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.466943424 +0000 UTC m=+156.608854653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.063750 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" event={"ID":"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba","Type":"ContainerStarted","Data":"22bbc8e43a9ff7b93f6a89f7e95fef5a4a2ff844266c76557c6ad450c7e51a82"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.068290 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.068620 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.56860899 +0000 UTC m=+156.710520219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.079680 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-59lwv" event={"ID":"cfc3766b-745f-4fe8-b1e5-beb39019ab01","Type":"ContainerStarted","Data":"a0d3f6667ac90d012b7bfb6f9685d599058c20b06f18a71a3e7f3d14de4e815e"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.092915 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" event={"ID":"09e7e6a7-b460-4f2d-9c66-28b41381364b","Type":"ContainerStarted","Data":"5f6945d2b0e435b2fef53b4c8cf2662373edca00ef858172d1fd6688cbaa1c73"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.102186 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" event={"ID":"1a2c5a8d-266b-441a-9977-16b138ebcc6f","Type":"ContainerStarted","Data":"a503f87f46275060687617261477a76d0477648eeedf9015ff7f0d07d2f48977"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.102222 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" event={"ID":"1a2c5a8d-266b-441a-9977-16b138ebcc6f","Type":"ContainerStarted","Data":"b8969ac0ce41a706aab7a8d93b5b272a289b1a9c351d1dd138a415719c0d8d58"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.103814 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.105861 4990 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-25q7v container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.105891 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" podUID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.114727 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lsgzp" event={"ID":"383ca895-74a4-4dd3-a431-e0caadfa246c","Type":"ContainerStarted","Data":"76bf1c820bec5a2bbfec03f68a146e3dc9b61c4dabd727738a0398ac7f8d45f5"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.114771 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lsgzp" event={"ID":"383ca895-74a4-4dd3-a431-e0caadfa246c","Type":"ContainerStarted","Data":"7fc4fc3ec4552842c17977cd47164515297fb4d0ac70e92a58c9fcb2114fda3e"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.139719 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" event={"ID":"650a3b2c-a031-438d-90e2-2daec51b80e1","Type":"ContainerStarted","Data":"8fceee0def4fb9d903ace449f719378b7e820989128eb6a282540afe4690509b"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.145125 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.152177 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" event={"ID":"f9060dac-4a34-4558-be5e-c16d356839b5","Type":"ContainerStarted","Data":"194c6d4157ee700591f704253cba036ceeebd7702a1cbc371d7b86eb678a22c3"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.183583 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.183630 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-89sr2" event={"ID":"25619895-5809-4fe0-baeb-1746885abadd","Type":"ContainerStarted","Data":"a32145af5a67b3f9ecc60c5b30b39ad59e0c3d5f3e134595f77b4b237c9bc3ef"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.183665 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-89sr2" event={"ID":"25619895-5809-4fe0-baeb-1746885abadd","Type":"ContainerStarted","Data":"2efdd9e4484c8cce7b52666826b16e8ff5a516f4cbf9f3ddfc6b6efdeec47b2d"} Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.185070 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.685051067 +0000 UTC m=+156.826962336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.185611 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.191095 4990 patch_prober.go:28] interesting pod/console-operator-58897d9998-89sr2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.191149 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-89sr2" podUID="25619895-5809-4fe0-baeb-1746885abadd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.217657 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" event={"ID":"96f89cc4-d412-486d-bc3d-f819b1537477","Type":"ContainerStarted","Data":"a37e43d1af5f4c5d787b1b56569db6044567209fab46bd1a8fc6f44987ef0f48"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.229999 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-fgc99" event={"ID":"21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c","Type":"ContainerStarted","Data":"58eec67968be51169ac5df6c3437b2757f5538c2796d282320828ba53091bd58"} Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.238821 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.238882 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.245169 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tk4fc" podStartSLOduration=134.245153585 podStartE2EDuration="2m14.245153585s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.243993585 +0000 UTC m=+156.385904824" watchObservedRunningTime="2025-12-03 12:40:08.245153585 +0000 UTC m=+156.387064814" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.286433 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.286677 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9cr2s" podStartSLOduration=134.286653603 podStartE2EDuration="2m14.286653603s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.282818714 +0000 UTC m=+156.424729953" watchObservedRunningTime="2025-12-03 12:40:08.286653603 +0000 UTC m=+156.428564832" Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.292423 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.792406451 +0000 UTC m=+156.934317690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.399237 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.399621 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.8995863 +0000 UTC m=+157.041497529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.399983 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.403045 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:08.903033999 +0000 UTC m=+157.044945228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.512959 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.513474 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.013454971 +0000 UTC m=+157.155366200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.535070 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jfgrs" podStartSLOduration=133.535052137 podStartE2EDuration="2m13.535052137s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.533936639 +0000 UTC m=+156.675847888" watchObservedRunningTime="2025-12-03 12:40:08.535052137 +0000 UTC m=+156.676963366" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.615682 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.616867 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.116852502 +0000 UTC m=+157.258763731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.716511 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.716811 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.216792915 +0000 UTC m=+157.358704144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.716994 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.717554 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.217544265 +0000 UTC m=+157.359455494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.744205 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" podStartSLOduration=134.74418175 podStartE2EDuration="2m14.74418175s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.716554309 +0000 UTC m=+156.858465558" watchObservedRunningTime="2025-12-03 12:40:08.74418175 +0000 UTC m=+156.886092979" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.744472 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" podStartSLOduration=133.744463627 podStartE2EDuration="2m13.744463627s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.740230999 +0000 UTC m=+156.882142228" watchObservedRunningTime="2025-12-03 12:40:08.744463627 +0000 UTC m=+156.886374856" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.811305 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jkj9" podStartSLOduration=134.811289888 podStartE2EDuration="2m14.811289888s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.784295082 +0000 UTC m=+156.926206311" watchObservedRunningTime="2025-12-03 12:40:08.811289888 +0000 UTC m=+156.953201117" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.812587 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4"] Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.824225 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.824666 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.324651021 +0000 UTC m=+157.466562250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.860114 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8"] Dec 03 12:40:08 crc kubenswrapper[4990]: W1203 12:40:08.870367 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda36bdb31_4352_4ef2_9d56_85cd07b89e21.slice/crio-dac68e861f435572bb15c776e5569d566c09b6c61c128f9484b12a86d7dc30de WatchSource:0}: Error finding container dac68e861f435572bb15c776e5569d566c09b6c61c128f9484b12a86d7dc30de: Status 404 returned error can't find the container with id dac68e861f435572bb15c776e5569d566c09b6c61c128f9484b12a86d7dc30de Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.880978 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zdzf9"] Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.925806 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:08 crc kubenswrapper[4990]: E1203 12:40:08.926197 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.426185375 +0000 UTC m=+157.568096604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.934405 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gmsz9"] Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.953378 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xs2zs" podStartSLOduration=134.953361644 podStartE2EDuration="2m14.953361644s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.949063204 +0000 UTC m=+157.090974433" watchObservedRunningTime="2025-12-03 12:40:08.953361644 +0000 UTC m=+157.095272863" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.957611 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.981763 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-p8mmq"] Dec 03 12:40:08 crc kubenswrapper[4990]: I1203 12:40:08.983599 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.001448 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lsgzp" podStartSLOduration=6.001425181 podStartE2EDuration="6.001425181s" podCreationTimestamp="2025-12-03 12:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:08.981791896 +0000 UTC m=+157.123703135" watchObservedRunningTime="2025-12-03 12:40:09.001425181 +0000 UTC m=+157.143336410" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.026448 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.026940 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.526922038 +0000 UTC m=+157.668833267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.105858 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-89sr2" podStartSLOduration=135.105832739 podStartE2EDuration="2m15.105832739s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:09.092812624 +0000 UTC m=+157.234723863" watchObservedRunningTime="2025-12-03 12:40:09.105832739 +0000 UTC m=+157.247743998" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.129392 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.129805 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.629792236 +0000 UTC m=+157.771703465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.152055 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" podStartSLOduration=135.152040289 podStartE2EDuration="2m15.152040289s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:09.137975087 +0000 UTC m=+157.279886316" watchObservedRunningTime="2025-12-03 12:40:09.152040289 +0000 UTC m=+157.293951518" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.234667 4990 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dqrqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.234724 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.235161 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.235302 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.735281851 +0000 UTC m=+157.877193080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.235516 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.235823 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.735808445 +0000 UTC m=+157.877719674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.237878 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:09 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:09 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:09 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.237931 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.241570 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-fgc99" podStartSLOduration=134.241545933 podStartE2EDuration="2m14.241545933s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:09.235136637 +0000 UTC m=+157.377047886" watchObservedRunningTime="2025-12-03 12:40:09.241545933 +0000 UTC m=+157.383457172" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.281415 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" event={"ID":"96f89cc4-d412-486d-bc3d-f819b1537477","Type":"ContainerStarted","Data":"a69f8dc686f20ba16b58966510c7f25daa25b5df73f9dcf888a5387fe6ed2294"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.293165 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" event={"ID":"11fe8587-dd60-4149-853e-db89b1cae352","Type":"ContainerStarted","Data":"02e78a15bf71ccc8ad1c2fda1a7aaffa356d6e7523863c42918b8dee8b0d76b9"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.305452 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" event={"ID":"0fee873c-8deb-4f94-aa19-632e04acd7de","Type":"ContainerStarted","Data":"5ecee368a7e5e1aef6b154adcaad8a999d1d09b53d2450c72faf396c6785937e"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.329495 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.347376 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.347820 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.847798948 +0000 UTC m=+157.989710177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.445492 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.448489 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" event={"ID":"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc","Type":"ContainerStarted","Data":"4578844277f9d4e04755ea391879b28c43587a85432ce2d4138aaf79ac3117ec"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.456391 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.456726 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:09.956713161 +0000 UTC m=+158.098624390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.457639 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" event={"ID":"7c2e0256-7681-49d7-afbc-145335bf5b17","Type":"ContainerStarted","Data":"a26d15c1892b9c17239b653a20a31068cff4ba7af5171bc6b075ae339560b79b"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.477979 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-59lwv" event={"ID":"cfc3766b-745f-4fe8-b1e5-beb39019ab01","Type":"ContainerStarted","Data":"d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.546424 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" event={"ID":"a36bdb31-4352-4ef2-9d56-85cd07b89e21","Type":"ContainerStarted","Data":"dac68e861f435572bb15c776e5569d566c09b6c61c128f9484b12a86d7dc30de"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.550918 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-59lwv" podStartSLOduration=135.550898735 podStartE2EDuration="2m15.550898735s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:09.533926029 +0000 UTC m=+157.675837278" watchObservedRunningTime="2025-12-03 12:40:09.550898735 +0000 UTC m=+157.692809964" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.551316 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.567333 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z8465" event={"ID":"bc1ccf16-c231-485b-9e14-8378f725292d","Type":"ContainerStarted","Data":"9dcdbae6f59240f7fd74b317a4c4f23fd806b038ee4c4b7358ed88ed35b3e219"} Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.570681 4990 patch_prober.go:28] interesting pod/console-operator-58897d9998-89sr2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.570725 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-89sr2" podUID="25619895-5809-4fe0-baeb-1746885abadd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.572312 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.572737 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.072713896 +0000 UTC m=+158.214625125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.587142 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.598963 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.600113 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.100055651 +0000 UTC m=+158.241966880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.646883 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" podStartSLOduration=134.646861605 podStartE2EDuration="2m14.646861605s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:09.636076028 +0000 UTC m=+157.777987277" watchObservedRunningTime="2025-12-03 12:40:09.646861605 +0000 UTC m=+157.788772834" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.689697 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.691473 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.191440083 +0000 UTC m=+158.333351312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.710367 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.748119 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.766213 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.774948 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.786784 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rftgm"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.788627 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.792490 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.792872 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.292859604 +0000 UTC m=+158.434770833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.804332 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.843977 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.867789 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.894536 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:09 crc kubenswrapper[4990]: E1203 12:40:09.894911 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.39489212 +0000 UTC m=+158.536803349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.915999 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.939363 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bvfbd"] Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.966836 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:09 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:09 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:09 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:09 crc kubenswrapper[4990]: I1203 12:40:09.966876 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.019317 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.019674 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.519657472 +0000 UTC m=+158.661568701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.120188 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.120532 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.620503868 +0000 UTC m=+158.762415097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.223512 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.224004 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.723988921 +0000 UTC m=+158.865900160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: W1203 12:40:10.236534 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod431416aa_9fd4_4c6a_8668_9e86d8c256aa.slice/crio-ec5b6087d6696d1cdf5dbd5d4e2fb80cb68f69af8dd33485df089fc8b64472d2 WatchSource:0}: Error finding container ec5b6087d6696d1cdf5dbd5d4e2fb80cb68f69af8dd33485df089fc8b64472d2: Status 404 returned error can't find the container with id ec5b6087d6696d1cdf5dbd5d4e2fb80cb68f69af8dd33485df089fc8b64472d2 Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.349861 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.351464 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.851440982 +0000 UTC m=+158.993352211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.454352 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.454720 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:10.95470597 +0000 UTC m=+159.096617199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.600338 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.600822 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.10080045 +0000 UTC m=+159.242711679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.634305 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" event={"ID":"7c2e0256-7681-49d7-afbc-145335bf5b17","Type":"ContainerStarted","Data":"3fbb1effa1020dd7baaa9926793a3c029ec00fdb6e59707a91989c18cb4d01a7"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.646931 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" event={"ID":"2d608434-0719-4924-9a9d-7f03de7b6cde","Type":"ContainerStarted","Data":"acfb2f6c0a2fd78aadd0163c3ae0d706e459712915e6832591a259d000fcd058"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.655389 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" event={"ID":"bdc9a406-d80a-4122-b556-5e0c1a12177b","Type":"ContainerStarted","Data":"093f1075ba3d5f8be09c16a6b7137a930e905f3ff9817cbcd0097b4d910a2995"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.664902 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" event={"ID":"11fe8587-dd60-4149-853e-db89b1cae352","Type":"ContainerStarted","Data":"67545df3f72adc276846ba62b7beaf456d01038170290ad4598b61fb1f5ce010"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.679289 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" event={"ID":"7b8daa8e-d841-4fd6-b34d-74f051da71eb","Type":"ContainerStarted","Data":"2142076c2625380ad5d3d04565f6a67c93ec76664d295ce4d6d2590b78ecfe18"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.687997 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" event={"ID":"5fc9eb76-3dd5-49ed-be37-a5923f4de448","Type":"ContainerStarted","Data":"a51f475e257817752c83af0c90b1f6bf206802aff93e0380ec498c6435e6ce63"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.688055 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" event={"ID":"5fc9eb76-3dd5-49ed-be37-a5923f4de448","Type":"ContainerStarted","Data":"5967ec8094b28792d0f4319d239305997e054737c7550b68bcec135b596ad926"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.689077 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.690650 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" event={"ID":"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba","Type":"ContainerStarted","Data":"627f9bfc8b14b3a3a74bfc0212aae940a096334df9fd0f0707a80994bd3e2d5f"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.691324 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" event={"ID":"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff","Type":"ContainerStarted","Data":"569015a08527f9e9b6ff18a5933769befe1c3b050ed140df8b546b1e2353d29c"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.691959 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" event={"ID":"cb47fbea-eda2-4b31-b64f-424cd42abf4e","Type":"ContainerStarted","Data":"9b4e3b8b45f48f7a146e637323b2febc6af4658db92e1a6a7672254ebdea4432"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.692608 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" event={"ID":"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe","Type":"ContainerStarted","Data":"ebeed61631000fdea98c81487c6997d80f3a44e0009582d62f838bd7e2d2da6d"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.693408 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jsjk4" event={"ID":"a36bdb31-4352-4ef2-9d56-85cd07b89e21","Type":"ContainerStarted","Data":"1922a4e35634ad0392d60c5134e456cfbef63905eacf1ceefa41087ce0e43b4f"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.696949 4990 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8ctlj container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.697015 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" podUID="5fc9eb76-3dd5-49ed-be37-a5923f4de448" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.704674 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bvfbd" event={"ID":"e8387a72-f04e-4004-82c8-5634a54600af","Type":"ContainerStarted","Data":"bf0f0fdcb26452dd3b49f6aaf93fb4f53d66d38cfdc44eab9273fe44898f1263"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.704860 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.705287 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.205274979 +0000 UTC m=+159.347186208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.713777 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" event={"ID":"431416aa-9fd4-4c6a-8668-9e86d8c256aa","Type":"ContainerStarted","Data":"ec5b6087d6696d1cdf5dbd5d4e2fb80cb68f69af8dd33485df089fc8b64472d2"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.714576 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" event={"ID":"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f","Type":"ContainerStarted","Data":"0905ecffc6a47042e35954f64080bd0ca8aaa5e918b72daba1ad930866b01a0e"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.716987 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" event={"ID":"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c","Type":"ContainerStarted","Data":"5afd044869d467663838cb12bfa0a379014f7fb1bfbee62616739b171ce58562"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.717836 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" event={"ID":"0e245e7d-f3ac-46c3-9186-e7a8fffc02ad","Type":"ContainerStarted","Data":"1414bb6972f1c7bf8003475b30c7ef496b45ed9a19c1e31910f29927ad52cdfc"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.720914 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" event={"ID":"a1f76c88-6b38-4c91-a6db-aa9e35e4672a","Type":"ContainerStarted","Data":"152c184e3bad6edbaec668136d08d074809553c67487124d80080df10ff87ca6"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.720943 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" event={"ID":"a1f76c88-6b38-4c91-a6db-aa9e35e4672a","Type":"ContainerStarted","Data":"9ddbdcb3b530b25c93ce6d07997b861a64f9db659cb767f39deaaf23f7382a16"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.721991 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerStarted","Data":"dad46f22a4c84e3cb5e727b7b026fcc963fb293733d186a1db88d671df532985"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.728046 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerStarted","Data":"ec756ac85c90651c5bab3988b6f4ea7975dee2e4fa0a3807209cd2cbce049a7d"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.728111 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.735930 4990 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gmsz9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.736052 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.767662 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7vjl9" podStartSLOduration=136.763443707 podStartE2EDuration="2m16.763443707s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:10.703984596 +0000 UTC m=+158.845895815" watchObservedRunningTime="2025-12-03 12:40:10.763443707 +0000 UTC m=+158.905354936" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.771174 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-p8mmq" event={"ID":"fbdfce6d-1752-4940-a0e5-ee5bfa89d733","Type":"ContainerStarted","Data":"82e567d4e89db2e5f02a0e4eeb66d55107cd98b574b5c669a1a5764d23bff7dc"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.771207 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-p8mmq" event={"ID":"fbdfce6d-1752-4940-a0e5-ee5bfa89d733","Type":"ContainerStarted","Data":"66dfdd20d99f0608f5043f28d86297c02f942d7e927960271dc4c93843802ba6"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.809301 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.810741 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.310714504 +0000 UTC m=+159.452625733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.874143 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" event={"ID":"8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc","Type":"ContainerStarted","Data":"efa4508ff06556154d9946608725d7cde3fcaa3f60c69fe54a35cbc7f0db5475"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.893499 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podStartSLOduration=135.893463964 podStartE2EDuration="2m15.893463964s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:10.827324982 +0000 UTC m=+158.969236201" watchObservedRunningTime="2025-12-03 12:40:10.893463964 +0000 UTC m=+159.035375193" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.894115 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" podStartSLOduration=135.89410797 podStartE2EDuration="2m15.89410797s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:10.770741765 +0000 UTC m=+158.912652994" watchObservedRunningTime="2025-12-03 12:40:10.89410797 +0000 UTC m=+159.036019199" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.906477 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" event={"ID":"826bd110-77e7-484c-bd48-d95fa1b4a0cd","Type":"ContainerStarted","Data":"1e744ce6621e5c6f19b5929e389abe31ca26974e1dc45912e4b8a7846fa9bac1"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.906968 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:10 crc kubenswrapper[4990]: E1203 12:40:10.909483 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.409452955 +0000 UTC m=+159.551364184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.914149 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" event={"ID":"09e7e6a7-b460-4f2d-9c66-28b41381364b","Type":"ContainerStarted","Data":"d700cfb5ec581c2e30ace383da24ed2181f9f6fed393ef2ebec8db36430ae02b"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.920336 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-56ffp" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.932033 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lthk" podStartSLOduration=135.932013115 podStartE2EDuration="2m15.932013115s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:10.919276278 +0000 UTC m=+159.061187517" watchObservedRunningTime="2025-12-03 12:40:10.932013115 +0000 UTC m=+159.073924344" Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.953503 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" event={"ID":"0fee873c-8deb-4f94-aa19-632e04acd7de","Type":"ContainerStarted","Data":"96a402f400bdbce3b5a673f463cc428bf174b3cbcbba3e3608ced1eacd73f8d9"} Dec 03 12:40:10 crc kubenswrapper[4990]: I1203 12:40:10.976688 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" event={"ID":"dcf4e387-1c02-4274-a26c-c7ae9085f24f","Type":"ContainerStarted","Data":"e706a7057c445bbdaf01ad338e7a37f90b1c34ceb2f0f70bb7715e37bcc4a4c3"} Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.004223 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:11 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:11 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:11 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.004295 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.006605 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" podStartSLOduration=137.006590946 podStartE2EDuration="2m17.006590946s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:11.004685326 +0000 UTC m=+159.146596575" watchObservedRunningTime="2025-12-03 12:40:11.006590946 +0000 UTC m=+159.148502175" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.011796 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.012965 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.512944739 +0000 UTC m=+159.654855968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.082101 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-t2dw4" podStartSLOduration=137.082085868 podStartE2EDuration="2m17.082085868s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:11.081256418 +0000 UTC m=+159.223167657" watchObservedRunningTime="2025-12-03 12:40:11.082085868 +0000 UTC m=+159.223997107" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.122899 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.125392 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.625374743 +0000 UTC m=+159.767286052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.201000 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-p8mmq" podStartSLOduration=8.200982639 podStartE2EDuration="8.200982639s" podCreationTimestamp="2025-12-03 12:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:11.198786823 +0000 UTC m=+159.340698052" watchObservedRunningTime="2025-12-03 12:40:11.200982639 +0000 UTC m=+159.342893868" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.231272 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.231710 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.731690209 +0000 UTC m=+159.873601438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.333154 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.333496 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.83347989 +0000 UTC m=+159.975391119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.434375 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.434602 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.934569932 +0000 UTC m=+160.076481171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.434703 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.434974 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:11.934961922 +0000 UTC m=+160.076873151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.536260 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.536395 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.036357052 +0000 UTC m=+160.178268281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.536602 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.536947 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.036919716 +0000 UTC m=+160.178830955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.656068 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.656485 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.156418893 +0000 UTC m=+160.298330112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.665266 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5f4j8" podStartSLOduration=136.66524281 podStartE2EDuration="2m16.66524281s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:11.251940821 +0000 UTC m=+159.393852050" watchObservedRunningTime="2025-12-03 12:40:11.66524281 +0000 UTC m=+159.807154039" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.665771 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-55tzf" podStartSLOduration=136.665763543 podStartE2EDuration="2m16.665763543s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:11.665210269 +0000 UTC m=+159.807121498" watchObservedRunningTime="2025-12-03 12:40:11.665763543 +0000 UTC m=+159.807674782" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.765819 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.766241 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.266228679 +0000 UTC m=+160.408139918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.869341 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.874149 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.374118886 +0000 UTC m=+160.516030185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.962947 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:11 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:11 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:11 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.963026 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:11 crc kubenswrapper[4990]: I1203 12:40:11.973359 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:11 crc kubenswrapper[4990]: E1203 12:40:11.973743 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.47373084 +0000 UTC m=+160.615642069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.077211 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.077856 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.57783731 +0000 UTC m=+160.719748539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.167830 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" event={"ID":"acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f","Type":"ContainerStarted","Data":"5152abb47ee3bf8a2fad8ddcdc48443cc5a9d60bc0a502c9f1b0fe7e755c0f9b"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.169195 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.170579 4990 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p2clc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.170739 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" podUID="acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.181701 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.183255 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.683219693 +0000 UTC m=+160.825130912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.210646 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" event={"ID":"bdc9a406-d80a-4122-b556-5e0c1a12177b","Type":"ContainerStarted","Data":"f71eea2da4838c7fb5e423f4a3b7211dd86b3dbb790fb81b30ecea11aa578667"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.219901 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" event={"ID":"214b5d9c-7933-4ec3-b2ae-60ed6d752cfe","Type":"ContainerStarted","Data":"49154bc3efd713a5fc44c8c81eff774aed907709898e2fd20a7e50c1020ccaef"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.237160 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" podStartSLOduration=137.23714016 podStartE2EDuration="2m17.23714016s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:12.235807167 +0000 UTC m=+160.377718396" watchObservedRunningTime="2025-12-03 12:40:12.23714016 +0000 UTC m=+160.379051389" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.238725 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" event={"ID":"dcf4e387-1c02-4274-a26c-c7ae9085f24f","Type":"ContainerStarted","Data":"06dcba21618b5d5b2340b1e565838739b1b1d18f2bd9bbc57a4be23d595541f3"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.240645 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" event={"ID":"431416aa-9fd4-4c6a-8668-9e86d8c256aa","Type":"ContainerStarted","Data":"05a0e0ff0380b0d3a218f574f022ec53c8e09912fdbd917a8c699c82850b360e"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.241817 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.250451 4990 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pk52k container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.250537 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" podUID="431416aa-9fd4-4c6a-8668-9e86d8c256aa" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.258118 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" event={"ID":"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c","Type":"ContainerStarted","Data":"b21a7f33ca07b819c9ee25213769897fa0d43683512c5bf6c9f5e23fafdfddae"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.282749 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.283422 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.783408271 +0000 UTC m=+160.925319500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.322707 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" event={"ID":"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff","Type":"ContainerStarted","Data":"c160c1c15dbc244ac8b3e5ddc6b9c4a912d387939916ba13b510f62a833a984e"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.366558 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" event={"ID":"cb47fbea-eda2-4b31-b64f-424cd42abf4e","Type":"ContainerStarted","Data":"46e631129619eaca44c59be44bc65b03f76c0af1f7884b9351c900f982691683"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.379748 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" event={"ID":"d3bb2b9b-642c-4bf7-b7c8-401d13ac7eba","Type":"ContainerStarted","Data":"0e5ca5bc81e7417058de18b601b348cb192e0884d7cbf1cb7799111731aca57d"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.389442 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" event={"ID":"2d608434-0719-4924-9a9d-7f03de7b6cde","Type":"ContainerStarted","Data":"049ac3bc48544671e4491bf44be92d80bf010983f1f8db36de3d04c2f1d97b76"} Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.387399 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.887382228 +0000 UTC m=+161.029293457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.387063 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.391729 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bvfbd" event={"ID":"e8387a72-f04e-4004-82c8-5634a54600af","Type":"ContainerStarted","Data":"fe39b8ae5b4667799d106591e4587c58cb28b1672fa782a8d5c21e959d10cc2d"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.428142 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" event={"ID":"7b8daa8e-d841-4fd6-b34d-74f051da71eb","Type":"ContainerStarted","Data":"69bfc434c8064da489315ba44801157aeba72da213d737f25fc20cc2ebc82b38"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.442664 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rftgm" podStartSLOduration=137.44263859 podStartE2EDuration="2m17.44263859s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:12.277845239 +0000 UTC m=+160.419756488" watchObservedRunningTime="2025-12-03 12:40:12.44263859 +0000 UTC m=+160.584549839" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.493076 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.496530 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:12.996512657 +0000 UTC m=+161.138423886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.508745 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" event={"ID":"826bd110-77e7-484c-bd48-d95fa1b4a0cd","Type":"ContainerStarted","Data":"9b9832b269b2c2b671caf9164566a394f62c930840b8d2fde81058e9e2ed406c"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.514936 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" podStartSLOduration=137.514921241 podStartE2EDuration="2m17.514921241s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:12.513911775 +0000 UTC m=+160.655823004" watchObservedRunningTime="2025-12-03 12:40:12.514921241 +0000 UTC m=+160.656832470" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.525321 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z8465" event={"ID":"bc1ccf16-c231-485b-9e14-8378f725292d","Type":"ContainerStarted","Data":"6c424a3da1a4f1aa96c1718cd246091ce6422c976ec91e4ed238df00dd63bba8"} Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.527204 4990 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8ctlj container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.527250 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" podUID="5fc9eb76-3dd5-49ed-be37-a5923f4de448" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.527978 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.531940 4990 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gmsz9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.531999 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.596233 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.599841 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.099826646 +0000 UTC m=+161.241737875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.719314 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.719801 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.219784204 +0000 UTC m=+161.361695433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.820776 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.821416 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.3214044 +0000 UTC m=+161.463315629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:12 crc kubenswrapper[4990]: I1203 12:40:12.923064 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:12 crc kubenswrapper[4990]: E1203 12:40:12.923521 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.423500978 +0000 UTC m=+161.565412207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.020550 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:13 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:13 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:13 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.020615 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.026262 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.026579 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.52656754 +0000 UTC m=+161.668478769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.133251 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.133390 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.63337463 +0000 UTC m=+161.775285859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.133741 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.134090 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.634081008 +0000 UTC m=+161.775992237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.223075 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tp6lf" podStartSLOduration=138.223055708 podStartE2EDuration="2m18.223055708s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.221068377 +0000 UTC m=+161.362979606" watchObservedRunningTime="2025-12-03 12:40:13.223055708 +0000 UTC m=+161.364966937" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.234922 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.235198 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.73518266 +0000 UTC m=+161.877093889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.235260 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.235628 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.735617992 +0000 UTC m=+161.877529221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.362407 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vb5gm" podStartSLOduration=138.362391605 podStartE2EDuration="2m18.362391605s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.266520778 +0000 UTC m=+161.408432007" watchObservedRunningTime="2025-12-03 12:40:13.362391605 +0000 UTC m=+161.504302824" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.365105 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.369398 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" podStartSLOduration=139.369379415 podStartE2EDuration="2m19.369379415s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.361779279 +0000 UTC m=+161.503690518" watchObservedRunningTime="2025-12-03 12:40:13.369379415 +0000 UTC m=+161.511290644" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.370156 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.870135644 +0000 UTC m=+162.012046873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.471942 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.472517 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:13.972498699 +0000 UTC m=+162.114409928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.494663 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-h7z6x" podStartSLOduration=139.4946421 podStartE2EDuration="2m19.4946421s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.490884902 +0000 UTC m=+161.632796131" watchObservedRunningTime="2025-12-03 12:40:13.4946421 +0000 UTC m=+161.636553329" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.590991 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.591261 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.091233146 +0000 UTC m=+162.233144425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.599656 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" event={"ID":"bdc9a406-d80a-4122-b556-5e0c1a12177b","Type":"ContainerStarted","Data":"651e223d1b653141db8be9e0f742b0a4a40a61584977790200dd7ca576969a2c"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.649323 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" event={"ID":"a1f76c88-6b38-4c91-a6db-aa9e35e4672a","Type":"ContainerStarted","Data":"31f70da719b13e5d77c803ae039147d0d58d49538d00a7e1086cfc17fecdec53"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.651642 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" event={"ID":"dcf4e387-1c02-4274-a26c-c7ae9085f24f","Type":"ContainerStarted","Data":"85724215fec89751962c168e9897545995bb697b2bf7840454ff9eafd93b47ac"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.652359 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.654700 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bvfbd" event={"ID":"e8387a72-f04e-4004-82c8-5634a54600af","Type":"ContainerStarted","Data":"84892559339e68ed1eb9df39a20356f488a0715e8f0f31f5077ad96c25baf8a7"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.654730 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.662882 4990 generic.go:334] "Generic (PLEG): container finished" podID="7b8daa8e-d841-4fd6-b34d-74f051da71eb" containerID="69bfc434c8064da489315ba44801157aeba72da213d737f25fc20cc2ebc82b38" exitCode=0 Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.663439 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" event={"ID":"7b8daa8e-d841-4fd6-b34d-74f051da71eb","Type":"ContainerDied","Data":"69bfc434c8064da489315ba44801157aeba72da213d737f25fc20cc2ebc82b38"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.663493 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" event={"ID":"7b8daa8e-d841-4fd6-b34d-74f051da71eb","Type":"ContainerStarted","Data":"32e50e62b3058c28314547a0d448b43e0286aa5f1b744ec6b7801947e03aab32"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.665879 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" event={"ID":"7c2e0256-7681-49d7-afbc-145335bf5b17","Type":"ContainerStarted","Data":"f0826548c980a69ab3935c8ba01d8378cb61f4333d084cfc15e1809aa0729653"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.667658 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" event={"ID":"a1fb2e05-1e50-4d85-baa7-fb6dc003b53c","Type":"ContainerStarted","Data":"f619fedc618e6d1fec32afe4cfb8e17ce46841684dabdfe8fb808d70b1218d72"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.684809 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" event={"ID":"ef0fa8d8-e755-474d-bdf2-2f742d4a87ff","Type":"ContainerStarted","Data":"0b34e806bacd240958d9262011b097cd29064da4aaee805e5d98fe390bc89d02"} Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.684941 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.688933 4990 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pk52k container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.688971 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" podUID="431416aa-9fd4-4c6a-8668-9e86d8c256aa" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.689402 4990 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gmsz9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.689424 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.689489 4990 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8ctlj container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.689506 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" podUID="5fc9eb76-3dd5-49ed-be37-a5923f4de448" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.689543 4990 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p2clc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.689556 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" podUID="acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.691631 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6c8ms" podStartSLOduration=138.691620249 podStartE2EDuration="2m18.691620249s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.689899455 +0000 UTC m=+161.831810684" watchObservedRunningTime="2025-12-03 12:40:13.691620249 +0000 UTC m=+161.833531478" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.693248 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.693591 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.19357133 +0000 UTC m=+162.335482559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.746045 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fqfvb" podStartSLOduration=138.74602642 podStartE2EDuration="2m18.74602642s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.744806128 +0000 UTC m=+161.886717357" watchObservedRunningTime="2025-12-03 12:40:13.74602642 +0000 UTC m=+161.887937649" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.796002 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.797139 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.297125536 +0000 UTC m=+162.439036765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.825980 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" podStartSLOduration=138.825875835 podStartE2EDuration="2m18.825875835s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.823026912 +0000 UTC m=+161.964938141" watchObservedRunningTime="2025-12-03 12:40:13.825875835 +0000 UTC m=+161.967787064" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.828461 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-zdzf9" podStartSLOduration=138.828445002 podStartE2EDuration="2m18.828445002s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.794406316 +0000 UTC m=+161.936317545" watchObservedRunningTime="2025-12-03 12:40:13.828445002 +0000 UTC m=+161.970356231" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.897671 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:13 crc kubenswrapper[4990]: E1203 12:40:13.897978 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.397965951 +0000 UTC m=+162.539877180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.899076 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kqf6f" podStartSLOduration=138.899066119 podStartE2EDuration="2m18.899066119s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.851618859 +0000 UTC m=+161.993530088" watchObservedRunningTime="2025-12-03 12:40:13.899066119 +0000 UTC m=+162.040977348" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.958889 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bvfbd" podStartSLOduration=10.958870349 podStartE2EDuration="10.958870349s" podCreationTimestamp="2025-12-03 12:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.95583104 +0000 UTC m=+162.097742279" watchObservedRunningTime="2025-12-03 12:40:13.958870349 +0000 UTC m=+162.100781578" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.960619 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:13 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:13 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:13 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.960878 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:13 crc kubenswrapper[4990]: I1203 12:40:13.989378 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" podStartSLOduration=138.989362104 podStartE2EDuration="2m18.989362104s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:13.987878835 +0000 UTC m=+162.129790054" watchObservedRunningTime="2025-12-03 12:40:13.989362104 +0000 UTC m=+162.131273333" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.003390 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.004177 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.504153714 +0000 UTC m=+162.646064943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.016477 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdq68" podStartSLOduration=139.01644443 podStartE2EDuration="2m19.01644443s" podCreationTimestamp="2025-12-03 12:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:14.014793268 +0000 UTC m=+162.156704497" watchObservedRunningTime="2025-12-03 12:40:14.01644443 +0000 UTC m=+162.158355659" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.105579 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.105878 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.605866862 +0000 UTC m=+162.747778091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.206289 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.206530 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.706514783 +0000 UTC m=+162.848426012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.307970 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.308285 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.808272433 +0000 UTC m=+162.950183662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.424055 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.424206 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.924179585 +0000 UTC m=+163.066090814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.424716 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.425007 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:14.924995327 +0000 UTC m=+163.066906556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.525531 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.525703 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.025678579 +0000 UTC m=+163.167589808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.626886 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.627300 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.127285605 +0000 UTC m=+163.269196834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.728151 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.728428 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.228406687 +0000 UTC m=+163.370317916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.728505 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.728810 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.228797067 +0000 UTC m=+163.370708296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.740833 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z8465" event={"ID":"bc1ccf16-c231-485b-9e14-8378f725292d","Type":"ContainerStarted","Data":"095be7ebaf15b27b9f5b873a9b8efac1f558f8e2a781252ed4e4f7dc853e7d78"} Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.742274 4990 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p2clc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.742313 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" podUID="acdeba4a-d9ed-4a94-b447-7a1cf3e8c73f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.742407 4990 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pk52k container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.742422 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" podUID="431416aa-9fd4-4c6a-8668-9e86d8c256aa" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.831500 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.831628 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.331606043 +0000 UTC m=+163.473517282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.832674 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.835072 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.335051682 +0000 UTC m=+163.476962991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.928929 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.929252 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.929033 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.929308 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.933841 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.933978 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.433955288 +0000 UTC m=+163.575866517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.934136 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:14 crc kubenswrapper[4990]: E1203 12:40:14.934540 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.434531733 +0000 UTC m=+163.576442962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.939180 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.939227 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.993225 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:14 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:14 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:14 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:14 crc kubenswrapper[4990]: I1203 12:40:14.993303 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.036229 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.036326 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.536306003 +0000 UTC m=+163.678217232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.036896 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.038731 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.538720654 +0000 UTC m=+163.680631963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.138092 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.138525 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.638511064 +0000 UTC m=+163.780422293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.240139 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.240555 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.74053911 +0000 UTC m=+163.882450339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.341517 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.341702 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.841673213 +0000 UTC m=+163.983584442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.341866 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.342212 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.842201706 +0000 UTC m=+163.984112935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.443212 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.443586 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:15.943563636 +0000 UTC m=+164.085474865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.544409 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.544804 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.044786841 +0000 UTC m=+164.186698140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.559355 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-89sr2" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.562485 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.562518 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.563733 4990 patch_prober.go:28] interesting pod/console-f9d7485db-59lwv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.563772 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-59lwv" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.656242 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.666375 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.16634955 +0000 UTC m=+164.308260849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.750609 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z8465" event={"ID":"bc1ccf16-c231-485b-9e14-8378f725292d","Type":"ContainerStarted","Data":"4231e0f18fe3c31f136b11ff7a0a352dd0f29c9f930b2b15012b033e5a7a7f99"} Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.750668 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z8465" event={"ID":"bc1ccf16-c231-485b-9e14-8378f725292d","Type":"ContainerStarted","Data":"a7e3406bbcbbd11abeae96cda8e3baf171a8f376e2fd73feb10d934463324659"} Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.758382 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.758711 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.258699687 +0000 UTC m=+164.400610916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.773067 4990 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.792291 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-z8465" podStartSLOduration=12.792273572 podStartE2EDuration="12.792273572s" podCreationTimestamp="2025-12-03 12:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:15.788942625 +0000 UTC m=+163.930853854" watchObservedRunningTime="2025-12-03 12:40:15.792273572 +0000 UTC m=+163.934184791" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.859749 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.861268 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.361233737 +0000 UTC m=+164.503144966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.941262 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.941316 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.956413 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.959829 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:15 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:15 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:15 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.959918 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:15 crc kubenswrapper[4990]: I1203 12:40:15.961139 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:15 crc kubenswrapper[4990]: E1203 12:40:15.961870 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.461856377 +0000 UTC m=+164.603767696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.043795 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.044672 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.046761 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.050326 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.059771 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.061776 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:16 crc kubenswrapper[4990]: E1203 12:40:16.061947 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.561924032 +0000 UTC m=+164.703835261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.062036 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.062063 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.062232 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:16 crc kubenswrapper[4990]: E1203 12:40:16.062643 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.56262858 +0000 UTC m=+164.704539809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.120026 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.163988 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:16 crc kubenswrapper[4990]: E1203 12:40:16.164152 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.664117182 +0000 UTC m=+164.806028411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.164321 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.164349 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.164402 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.164628 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: E1203 12:40:16.164775 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 12:40:16.664767289 +0000 UTC m=+164.806678518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-twmj9" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.215791 4990 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T12:40:15.773280912Z","Handler":null,"Name":""} Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.232027 4990 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.232064 4990 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.255113 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.265929 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.347158 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8ctlj" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.359307 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p2clc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.373817 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.444963 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.477239 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.533211 4990 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.533318 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.648581 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk52k" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.661880 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j98rl"] Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.663176 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.666344 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.694505 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j98rl"] Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.781193 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-466jz\" (UniqueName: \"kubernetes.io/projected/89cb569a-c36f-481c-b5a0-00694046c00c-kube-api-access-466jz\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.781250 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-catalog-content\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.781291 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-utilities\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.827410 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-twmj9\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.862563 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jxn9m"] Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.864056 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.870054 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.881334 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jxn9m"] Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.882963 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jnqm\" (UniqueName: \"kubernetes.io/projected/8108de1d-75ea-419d-9658-e5a035088b36-kube-api-access-8jnqm\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883000 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-utilities\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883086 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-catalog-content\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883113 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-utilities\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883134 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-466jz\" (UniqueName: \"kubernetes.io/projected/89cb569a-c36f-481c-b5a0-00694046c00c-kube-api-access-466jz\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883155 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-catalog-content\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883618 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-catalog-content\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.883895 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-utilities\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.903341 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-466jz\" (UniqueName: \"kubernetes.io/projected/89cb569a-c36f-481c-b5a0-00694046c00c-kube-api-access-466jz\") pod \"community-operators-j98rl\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.962509 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:16 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:16 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:16 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.963145 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.984041 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-catalog-content\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.984101 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-utilities\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.984146 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jnqm\" (UniqueName: \"kubernetes.io/projected/8108de1d-75ea-419d-9658-e5a035088b36-kube-api-access-8jnqm\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.984947 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-catalog-content\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:16 crc kubenswrapper[4990]: I1203 12:40:16.986575 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.045421 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kjsgh"] Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.046985 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.059527 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kjsgh"] Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.067224 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.084922 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.093155 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a061fb3f-ace6-4ab6-a499-b34979186d08-metrics-certs\") pod \"network-metrics-daemon-6sqk8\" (UID: \"a061fb3f-ace6-4ab6-a499-b34979186d08\") " pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.186381 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-catalog-content\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.186446 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5pqb\" (UniqueName: \"kubernetes.io/projected/572c65be-7211-47d9-805e-081d07fbe844-kube-api-access-v5pqb\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.186511 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-utilities\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.204344 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6sqk8" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.240392 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tprnr"] Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.241626 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.256479 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tprnr"] Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.288291 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-utilities\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.288391 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-catalog-content\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.288418 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5pqb\" (UniqueName: \"kubernetes.io/projected/572c65be-7211-47d9-805e-081d07fbe844-kube-api-access-v5pqb\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.288861 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-catalog-content\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.289148 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-utilities\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.312951 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5pqb\" (UniqueName: \"kubernetes.io/projected/572c65be-7211-47d9-805e-081d07fbe844-kube-api-access-v5pqb\") pod \"community-operators-kjsgh\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.368341 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.389790 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-catalog-content\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.390317 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-utilities\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.390378 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc8t9\" (UniqueName: \"kubernetes.io/projected/ed381166-1062-4ca5-a553-81ab9681698e-kube-api-access-cc8t9\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.491268 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-utilities\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.491334 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc8t9\" (UniqueName: \"kubernetes.io/projected/ed381166-1062-4ca5-a553-81ab9681698e-kube-api-access-cc8t9\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.491385 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-catalog-content\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.492014 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-utilities\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.492112 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-catalog-content\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.510367 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc8t9\" (UniqueName: \"kubernetes.io/projected/ed381166-1062-4ca5-a553-81ab9681698e-kube-api-access-cc8t9\") pod \"certified-operators-tprnr\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.569097 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.959900 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:17 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:17 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:17 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:17 crc kubenswrapper[4990]: I1203 12:40:17.959970 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.273745 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.635183 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dbqqz"] Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.636283 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.639623 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.650278 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbqqz"] Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.705269 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zj55\" (UniqueName: \"kubernetes.io/projected/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-kube-api-access-5zj55\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.705521 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-utilities\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.705583 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-catalog-content\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.798313 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-utilities\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.803344 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jnqm\" (UniqueName: \"kubernetes.io/projected/8108de1d-75ea-419d-9658-e5a035088b36-kube-api-access-8jnqm\") pod \"certified-operators-jxn9m\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.815722 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zj55\" (UniqueName: \"kubernetes.io/projected/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-kube-api-access-5zj55\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.815773 4990 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8f7ss container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]log ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]etcd ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/max-in-flight-filter ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 12:40:18 crc kubenswrapper[4990]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 12:40:18 crc kubenswrapper[4990]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/openshift.io-startinformers ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 12:40:18 crc kubenswrapper[4990]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 12:40:18 crc kubenswrapper[4990]: livez check failed Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.815839 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" podUID="8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.815855 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-utilities\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.815911 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-catalog-content\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.816511 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-catalog-content\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.816582 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-utilities\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.873192 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.876635 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zj55\" (UniqueName: \"kubernetes.io/projected/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-kube-api-access-5zj55\") pod \"redhat-marketplace-dbqqz\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.930641 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:18 crc kubenswrapper[4990]: I1203 12:40:18.951137 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qxchk" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:18.956389 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:18.964501 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:19 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:19 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:19 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:18.964599 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:18.978926 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.245344 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cbxg8"] Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.246647 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.268533 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbxg8"] Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.350954 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-catalog-content\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.351049 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-utilities\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.351084 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd42h\" (UniqueName: \"kubernetes.io/projected/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-kube-api-access-jd42h\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.448348 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.449181 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.453045 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.453253 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd42h\" (UniqueName: \"kubernetes.io/projected/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-kube-api-access-jd42h\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.453350 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-catalog-content\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.453388 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.453422 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-utilities\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.454198 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-utilities\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.455415 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-catalog-content\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.471388 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.476988 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd42h\" (UniqueName: \"kubernetes.io/projected/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-kube-api-access-jd42h\") pod \"redhat-marketplace-cbxg8\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.559868 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ac5c2af-97e9-4e14-8489-4d87435d5856-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.560294 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ac5c2af-97e9-4e14-8489-4d87435d5856-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.639919 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.661979 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ac5c2af-97e9-4e14-8489-4d87435d5856-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.662051 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ac5c2af-97e9-4e14-8489-4d87435d5856-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.662153 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ac5c2af-97e9-4e14-8489-4d87435d5856-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.749880 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ac5c2af-97e9-4e14-8489-4d87435d5856-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.805790 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.806037 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"feff5a1f-1971-4743-aba5-46ae4c2aca8a","Type":"ContainerStarted","Data":"7c51cad5ef134ee5dfeb4bf08298f3eac23f884931deb5c9bde8c84dd38a30d6"} Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.835115 4990 generic.go:334] "Generic (PLEG): container finished" podID="826bd110-77e7-484c-bd48-d95fa1b4a0cd" containerID="9b9832b269b2c2b671caf9164566a394f62c930840b8d2fde81058e9e2ed406c" exitCode=0 Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.835529 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" event={"ID":"826bd110-77e7-484c-bd48-d95fa1b4a0cd","Type":"ContainerDied","Data":"9b9832b269b2c2b671caf9164566a394f62c930840b8d2fde81058e9e2ed406c"} Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.874489 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zn87g"] Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.875777 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.882860 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.907735 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zn87g"] Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.966730 4990 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8f7ss container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]log ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]etcd ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/max-in-flight-filter ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 12:40:19 crc kubenswrapper[4990]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/openshift.io-startinformers ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 12:40:19 crc kubenswrapper[4990]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 12:40:19 crc kubenswrapper[4990]: livez check failed Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.966787 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" podUID="8d96c8b7-b0dd-49ca-b8f4-c6ffad9d8ffc" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.967726 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:19 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:19 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:19 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.967781 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.973272 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nrdf\" (UniqueName: \"kubernetes.io/projected/34bb2446-2a47-4b35-b5a1-13d8de74822e-kube-api-access-5nrdf\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.973322 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-catalog-content\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:19 crc kubenswrapper[4990]: I1203 12:40:19.973369 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-utilities\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:19.997168 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6sqk8"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.077330 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nrdf\" (UniqueName: \"kubernetes.io/projected/34bb2446-2a47-4b35-b5a1-13d8de74822e-kube-api-access-5nrdf\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.077393 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-catalog-content\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.077426 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-utilities\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.077961 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-utilities\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.078584 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-catalog-content\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.104548 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nrdf\" (UniqueName: \"kubernetes.io/projected/34bb2446-2a47-4b35-b5a1-13d8de74822e-kube-api-access-5nrdf\") pod \"redhat-operators-zn87g\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.141027 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-twmj9"] Dec 03 12:40:20 crc kubenswrapper[4990]: W1203 12:40:20.181626 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod791d5adb_b1b5_4cf6_a003_f8ad41da334e.slice/crio-ea9073754fadc6a03a7bbb142eb2d380cfaca19c9f945daa6d63f44e4dec6b58 WatchSource:0}: Error finding container ea9073754fadc6a03a7bbb142eb2d380cfaca19c9f945daa6d63f44e4dec6b58: Status 404 returned error can't find the container with id ea9073754fadc6a03a7bbb142eb2d380cfaca19c9f945daa6d63f44e4dec6b58 Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.196908 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.256929 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sfvnj"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.258333 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.316128 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfvnj"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.359399 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tprnr"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.364864 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kjsgh"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.370842 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j98rl"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.388498 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tthkk\" (UniqueName: \"kubernetes.io/projected/2069abff-1d0a-42a0-8929-75acceecbe8b-kube-api-access-tthkk\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.388542 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-utilities\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.388615 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-catalog-content\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.393981 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jxn9m"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.473727 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbqqz"] Dec 03 12:40:20 crc kubenswrapper[4990]: W1203 12:40:20.485738 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddec4b1a_b34e_45f7_95e6_a04d29dde82c.slice/crio-7a83fc586282919f5e05615f3cc72d05a6d738424fe93e0be8f8107656db5e63 WatchSource:0}: Error finding container 7a83fc586282919f5e05615f3cc72d05a6d738424fe93e0be8f8107656db5e63: Status 404 returned error can't find the container with id 7a83fc586282919f5e05615f3cc72d05a6d738424fe93e0be8f8107656db5e63 Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.489906 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-catalog-content\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.489955 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tthkk\" (UniqueName: \"kubernetes.io/projected/2069abff-1d0a-42a0-8929-75acceecbe8b-kube-api-access-tthkk\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.489976 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-utilities\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.490336 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-utilities\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.490639 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-catalog-content\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.531367 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tthkk\" (UniqueName: \"kubernetes.io/projected/2069abff-1d0a-42a0-8929-75acceecbe8b-kube-api-access-tthkk\") pod \"redhat-operators-sfvnj\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.583288 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbxg8"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.623022 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.753320 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.788368 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zn87g"] Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.849685 4990 generic.go:334] "Generic (PLEG): container finished" podID="ed381166-1062-4ca5-a553-81ab9681698e" containerID="ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175" exitCode=0 Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.850079 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerDied","Data":"ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.850131 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerStarted","Data":"cef556ba78ee1616197a0dc73025df70cc0a2ccfa42b4613d52a025795f2b597"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.857866 4990 generic.go:334] "Generic (PLEG): container finished" podID="572c65be-7211-47d9-805e-081d07fbe844" containerID="d2c2ad962f16fbd9d4e23f30b5af1304d095931b50bcc1e86fbd5a5539af9056" exitCode=0 Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.858043 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerDied","Data":"d2c2ad962f16fbd9d4e23f30b5af1304d095931b50bcc1e86fbd5a5539af9056"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.858071 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerStarted","Data":"be6dae5a15c07a0ec8396c205ab2c0a26f5d2fdf7425803e82d3b194e9a68e58"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.860070 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ac5c2af-97e9-4e14-8489-4d87435d5856","Type":"ContainerStarted","Data":"f6246bcb489fe4337e8dfc42cc6377875cc0df7572ff1a6ed8acaa63041a5ae5"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.861284 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerStarted","Data":"f2eece2293c6748c775dc6db15f514023cd3c193ffc2180cec5c79f1f47b4539"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.861303 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerStarted","Data":"26d08bb85944788450f37a8b4d937de0db7a6b94fde9e0ce0ee3f3bd56961d28"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.863127 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" event={"ID":"a061fb3f-ace6-4ab6-a499-b34979186d08","Type":"ContainerStarted","Data":"0d18e03a67b561e22035068c8d7fb4f7316cecdb3e3ad4de3de7aa5461351436"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.864188 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" event={"ID":"a061fb3f-ace6-4ab6-a499-b34979186d08","Type":"ContainerStarted","Data":"869f4191ac3840451f8a623dd0ec57b0a8e887ae64fc39c5374443378de44d7c"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.884833 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"feff5a1f-1971-4743-aba5-46ae4c2aca8a","Type":"ContainerStarted","Data":"bd7580be9afe063991dfd3b3689382f1533e6917c0d66c3ab0836c2b1c3fa28e"} Dec 03 12:40:20 crc kubenswrapper[4990]: W1203 12:40:20.885725 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34bb2446_2a47_4b35_b5a1_13d8de74822e.slice/crio-4ccfd97112ed203840a201232e8b26b75883eca509b285caadab3e26606a8dbf WatchSource:0}: Error finding container 4ccfd97112ed203840a201232e8b26b75883eca509b285caadab3e26606a8dbf: Status 404 returned error can't find the container with id 4ccfd97112ed203840a201232e8b26b75883eca509b285caadab3e26606a8dbf Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.901822 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbxg8" event={"ID":"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030","Type":"ContainerStarted","Data":"70eeb3b465b52601a826e86e0e281394dfa0bce3dca6ddcc2aca1e580685d292"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.913002 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.955418 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" event={"ID":"791d5adb-b1b5-4cf6-a003-f8ad41da334e","Type":"ContainerStarted","Data":"50a28a3f7d4be32638a809532df45f870241910d4e0897677fe9664cd5100c0d"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.955475 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" event={"ID":"791d5adb-b1b5-4cf6-a003-f8ad41da334e","Type":"ContainerStarted","Data":"ea9073754fadc6a03a7bbb142eb2d380cfaca19c9f945daa6d63f44e4dec6b58"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.955974 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.964753 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:20 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:20 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:20 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.964803 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.970253 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbqqz" event={"ID":"ddec4b1a-b34e-45f7-95e6-a04d29dde82c","Type":"ContainerStarted","Data":"7a83fc586282919f5e05615f3cc72d05a6d738424fe93e0be8f8107656db5e63"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.974414 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.97438942 podStartE2EDuration="4.97438942s" podCreationTimestamp="2025-12-03 12:40:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:20.965361167 +0000 UTC m=+169.107272396" watchObservedRunningTime="2025-12-03 12:40:20.97438942 +0000 UTC m=+169.116300649" Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.977116 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerStarted","Data":"d8e3cd47f79a146ff14c7e46b1b5a45ffba08627a0aecb3a16e97cc100a5411d"} Dec 03 12:40:20 crc kubenswrapper[4990]: I1203 12:40:20.997702 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" podStartSLOduration=146.997680149 podStartE2EDuration="2m26.997680149s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:20.995481173 +0000 UTC m=+169.137392412" watchObservedRunningTime="2025-12-03 12:40:20.997680149 +0000 UTC m=+169.139591378" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.077598 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfvnj"] Dec 03 12:40:21 crc kubenswrapper[4990]: W1203 12:40:21.158880 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2069abff_1d0a_42a0_8929_75acceecbe8b.slice/crio-f873bd3fb51af49d95eb0e04d9ecd566d18fbb4a4561c303ec40ba6ae74776b2 WatchSource:0}: Error finding container f873bd3fb51af49d95eb0e04d9ecd566d18fbb4a4561c303ec40ba6ae74776b2: Status 404 returned error can't find the container with id f873bd3fb51af49d95eb0e04d9ecd566d18fbb4a4561c303ec40ba6ae74776b2 Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.402136 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.450178 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bvfbd" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.535379 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/826bd110-77e7-484c-bd48-d95fa1b4a0cd-secret-volume\") pod \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.535543 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/826bd110-77e7-484c-bd48-d95fa1b4a0cd-config-volume\") pod \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.535667 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmq6k\" (UniqueName: \"kubernetes.io/projected/826bd110-77e7-484c-bd48-d95fa1b4a0cd-kube-api-access-xmq6k\") pod \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\" (UID: \"826bd110-77e7-484c-bd48-d95fa1b4a0cd\") " Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.536631 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826bd110-77e7-484c-bd48-d95fa1b4a0cd-config-volume" (OuterVolumeSpecName: "config-volume") pod "826bd110-77e7-484c-bd48-d95fa1b4a0cd" (UID: "826bd110-77e7-484c-bd48-d95fa1b4a0cd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.542368 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826bd110-77e7-484c-bd48-d95fa1b4a0cd-kube-api-access-xmq6k" (OuterVolumeSpecName: "kube-api-access-xmq6k") pod "826bd110-77e7-484c-bd48-d95fa1b4a0cd" (UID: "826bd110-77e7-484c-bd48-d95fa1b4a0cd"). InnerVolumeSpecName "kube-api-access-xmq6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.543272 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826bd110-77e7-484c-bd48-d95fa1b4a0cd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "826bd110-77e7-484c-bd48-d95fa1b4a0cd" (UID: "826bd110-77e7-484c-bd48-d95fa1b4a0cd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.637438 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/826bd110-77e7-484c-bd48-d95fa1b4a0cd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.637529 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmq6k\" (UniqueName: \"kubernetes.io/projected/826bd110-77e7-484c-bd48-d95fa1b4a0cd-kube-api-access-xmq6k\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.637548 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/826bd110-77e7-484c-bd48-d95fa1b4a0cd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.961354 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:21 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:21 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:21 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.961438 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.987587 4990 generic.go:334] "Generic (PLEG): container finished" podID="8108de1d-75ea-419d-9658-e5a035088b36" containerID="f2eece2293c6748c775dc6db15f514023cd3c193ffc2180cec5c79f1f47b4539" exitCode=0 Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.987767 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerDied","Data":"f2eece2293c6748c775dc6db15f514023cd3c193ffc2180cec5c79f1f47b4539"} Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.990795 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerStarted","Data":"4ccfd97112ed203840a201232e8b26b75883eca509b285caadab3e26606a8dbf"} Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.992752 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerStarted","Data":"f873bd3fb51af49d95eb0e04d9ecd566d18fbb4a4561c303ec40ba6ae74776b2"} Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.996953 4990 generic.go:334] "Generic (PLEG): container finished" podID="89cb569a-c36f-481c-b5a0-00694046c00c" containerID="47e5ddb22e46cb429b080d91d979ec00a59f7a5d70be5f70030f50c5e950310e" exitCode=0 Dec 03 12:40:21 crc kubenswrapper[4990]: I1203 12:40:21.997376 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerDied","Data":"47e5ddb22e46cb429b080d91d979ec00a59f7a5d70be5f70030f50c5e950310e"} Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.000679 4990 generic.go:334] "Generic (PLEG): container finished" podID="feff5a1f-1971-4743-aba5-46ae4c2aca8a" containerID="bd7580be9afe063991dfd3b3689382f1533e6917c0d66c3ab0836c2b1c3fa28e" exitCode=0 Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.000769 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"feff5a1f-1971-4743-aba5-46ae4c2aca8a","Type":"ContainerDied","Data":"bd7580be9afe063991dfd3b3689382f1533e6917c0d66c3ab0836c2b1c3fa28e"} Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.007603 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.008535 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l" event={"ID":"826bd110-77e7-484c-bd48-d95fa1b4a0cd","Type":"ContainerDied","Data":"1e744ce6621e5c6f19b5929e389abe31ca26974e1dc45912e4b8a7846fa9bac1"} Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.008682 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e744ce6621e5c6f19b5929e389abe31ca26974e1dc45912e4b8a7846fa9bac1" Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.962944 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:22 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:22 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:22 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:22 crc kubenswrapper[4990]: I1203 12:40:22.963213 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.030190 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6sqk8" event={"ID":"a061fb3f-ace6-4ab6-a499-b34979186d08","Type":"ContainerStarted","Data":"9917eb35fde0a9638a9db67e914f7df3d304014b50730def67bbcc7e7e1e3bfb"} Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.034486 4990 generic.go:334] "Generic (PLEG): container finished" podID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerID="9935e4025e2531719e679337bceba3a8cc735f11fe2f938f2da6c76a4ac75962" exitCode=0 Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.034553 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerDied","Data":"9935e4025e2531719e679337bceba3a8cc735f11fe2f938f2da6c76a4ac75962"} Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.037344 4990 generic.go:334] "Generic (PLEG): container finished" podID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerID="dd143e7dfee45916b619826274787c7470faa2d2367ef9e0b0c0c17c776ef2d8" exitCode=0 Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.037388 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbxg8" event={"ID":"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030","Type":"ContainerDied","Data":"dd143e7dfee45916b619826274787c7470faa2d2367ef9e0b0c0c17c776ef2d8"} Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.043277 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ac5c2af-97e9-4e14-8489-4d87435d5856","Type":"ContainerStarted","Data":"23e18e885e2612379a982d9526be933d229cb7044204961da4378d84ec908882"} Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.045586 4990 generic.go:334] "Generic (PLEG): container finished" podID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerID="3f98eaa052dcdb6fa554f07427a51b5a33183160e271e246fb2e3578b2f0a653" exitCode=0 Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.045637 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerDied","Data":"3f98eaa052dcdb6fa554f07427a51b5a33183160e271e246fb2e3578b2f0a653"} Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.081298 4990 generic.go:334] "Generic (PLEG): container finished" podID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerID="ddb608ddbd922ae62062a0046c520b17134120f655d85c23fd62f4b391bb5145" exitCode=0 Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.082575 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbqqz" event={"ID":"ddec4b1a-b34e-45f7-95e6-a04d29dde82c","Type":"ContainerDied","Data":"ddb608ddbd922ae62062a0046c520b17134120f655d85c23fd62f4b391bb5145"} Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.101278 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-6sqk8" podStartSLOduration=149.101258567 podStartE2EDuration="2m29.101258567s" podCreationTimestamp="2025-12-03 12:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:23.056207667 +0000 UTC m=+171.198118896" watchObservedRunningTime="2025-12-03 12:40:23.101258567 +0000 UTC m=+171.243169806" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.183560 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.183542155 podStartE2EDuration="4.183542155s" podCreationTimestamp="2025-12-03 12:40:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:40:23.147317132 +0000 UTC m=+171.289228361" watchObservedRunningTime="2025-12-03 12:40:23.183542155 +0000 UTC m=+171.325453384" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.684197 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.827502 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kubelet-dir\") pod \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.827634 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "feff5a1f-1971-4743-aba5-46ae4c2aca8a" (UID: "feff5a1f-1971-4743-aba5-46ae4c2aca8a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.827660 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kube-api-access\") pod \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\" (UID: \"feff5a1f-1971-4743-aba5-46ae4c2aca8a\") " Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.828266 4990 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.837050 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "feff5a1f-1971-4743-aba5-46ae4c2aca8a" (UID: "feff5a1f-1971-4743-aba5-46ae4c2aca8a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.929547 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feff5a1f-1971-4743-aba5-46ae4c2aca8a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.959653 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:23 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:23 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:23 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:23 crc kubenswrapper[4990]: I1203 12:40:23.959724 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.103435 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"feff5a1f-1971-4743-aba5-46ae4c2aca8a","Type":"ContainerDied","Data":"7c51cad5ef134ee5dfeb4bf08298f3eac23f884931deb5c9bde8c84dd38a30d6"} Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.103500 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c51cad5ef134ee5dfeb4bf08298f3eac23f884931deb5c9bde8c84dd38a30d6" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.103507 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.936754 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.937101 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.936768 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.937556 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.944299 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.949653 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8f7ss" Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.960520 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:24 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:24 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:24 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:24 crc kubenswrapper[4990]: I1203 12:40:24.960581 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:25 crc kubenswrapper[4990]: I1203 12:40:25.213755 4990 generic.go:334] "Generic (PLEG): container finished" podID="6ac5c2af-97e9-4e14-8489-4d87435d5856" containerID="23e18e885e2612379a982d9526be933d229cb7044204961da4378d84ec908882" exitCode=0 Dec 03 12:40:25 crc kubenswrapper[4990]: I1203 12:40:25.213873 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ac5c2af-97e9-4e14-8489-4d87435d5856","Type":"ContainerDied","Data":"23e18e885e2612379a982d9526be933d229cb7044204961da4378d84ec908882"} Dec 03 12:40:25 crc kubenswrapper[4990]: I1203 12:40:25.577705 4990 patch_prober.go:28] interesting pod/console-f9d7485db-59lwv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 12:40:25 crc kubenswrapper[4990]: I1203 12:40:25.577779 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-59lwv" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 12:40:25 crc kubenswrapper[4990]: I1203 12:40:25.960393 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:25 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:25 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:25 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:25 crc kubenswrapper[4990]: I1203 12:40:25.960472 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:26 crc kubenswrapper[4990]: I1203 12:40:26.972133 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:26 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:26 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:26 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:26 crc kubenswrapper[4990]: I1203 12:40:26.972540 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.225437 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.365248 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ac5c2af-97e9-4e14-8489-4d87435d5856-kubelet-dir\") pod \"6ac5c2af-97e9-4e14-8489-4d87435d5856\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.365355 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ac5c2af-97e9-4e14-8489-4d87435d5856-kube-api-access\") pod \"6ac5c2af-97e9-4e14-8489-4d87435d5856\" (UID: \"6ac5c2af-97e9-4e14-8489-4d87435d5856\") " Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.365345 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ac5c2af-97e9-4e14-8489-4d87435d5856-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6ac5c2af-97e9-4e14-8489-4d87435d5856" (UID: "6ac5c2af-97e9-4e14-8489-4d87435d5856"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.365691 4990 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ac5c2af-97e9-4e14-8489-4d87435d5856-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.373792 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac5c2af-97e9-4e14-8489-4d87435d5856-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6ac5c2af-97e9-4e14-8489-4d87435d5856" (UID: "6ac5c2af-97e9-4e14-8489-4d87435d5856"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.440266 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ac5c2af-97e9-4e14-8489-4d87435d5856","Type":"ContainerDied","Data":"f6246bcb489fe4337e8dfc42cc6377875cc0df7572ff1a6ed8acaa63041a5ae5"} Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.440317 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6246bcb489fe4337e8dfc42cc6377875cc0df7572ff1a6ed8acaa63041a5ae5" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.440407 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.492372 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ac5c2af-97e9-4e14-8489-4d87435d5856-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.985234 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:27 crc kubenswrapper[4990]: [-]has-synced failed: reason withheld Dec 03 12:40:27 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:27 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:27 crc kubenswrapper[4990]: I1203 12:40:27.985324 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:28 crc kubenswrapper[4990]: I1203 12:40:28.965347 4990 patch_prober.go:28] interesting pod/router-default-5444994796-fgc99 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 12:40:28 crc kubenswrapper[4990]: [+]has-synced ok Dec 03 12:40:28 crc kubenswrapper[4990]: [+]process-running ok Dec 03 12:40:28 crc kubenswrapper[4990]: healthz check failed Dec 03 12:40:28 crc kubenswrapper[4990]: I1203 12:40:28.965720 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fgc99" podUID="21d9e1a4-fcc1-43a0-b18e-d8c068f48b2c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 12:40:29 crc kubenswrapper[4990]: I1203 12:40:29.963483 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:29 crc kubenswrapper[4990]: I1203 12:40:29.982352 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-fgc99" Dec 03 12:40:33 crc kubenswrapper[4990]: I1203 12:40:33.285859 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:40:33 crc kubenswrapper[4990]: I1203 12:40:33.286226 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.929718 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.929995 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.930466 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.930491 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.930522 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.931071 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"d639693cc9fc362114eaaf524724196a18b9e0b657416d2ed43f5cb1f4de098e"} pod="openshift-console/downloads-7954f5f757-lh8cf" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.931151 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" containerID="cri-o://d639693cc9fc362114eaaf524724196a18b9e0b657416d2ed43f5cb1f4de098e" gracePeriod=2 Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.931517 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:34 crc kubenswrapper[4990]: I1203 12:40:34.931540 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:35 crc kubenswrapper[4990]: I1203 12:40:35.608369 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:35 crc kubenswrapper[4990]: I1203 12:40:35.617282 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:40:35 crc kubenswrapper[4990]: I1203 12:40:35.893170 4990 generic.go:334] "Generic (PLEG): container finished" podID="80543db8-34e7-48cb-93db-897ae58288d2" containerID="d639693cc9fc362114eaaf524724196a18b9e0b657416d2ed43f5cb1f4de098e" exitCode=0 Dec 03 12:40:35 crc kubenswrapper[4990]: I1203 12:40:35.893399 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lh8cf" event={"ID":"80543db8-34e7-48cb-93db-897ae58288d2","Type":"ContainerDied","Data":"d639693cc9fc362114eaaf524724196a18b9e0b657416d2ed43f5cb1f4de098e"} Dec 03 12:40:37 crc kubenswrapper[4990]: I1203 12:40:37.096957 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:40:44 crc kubenswrapper[4990]: I1203 12:40:44.497022 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 12:40:44 crc kubenswrapper[4990]: I1203 12:40:44.940096 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:44 crc kubenswrapper[4990]: I1203 12:40:44.940177 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:46 crc kubenswrapper[4990]: I1203 12:40:46.398892 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2d2c8" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.835867 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 12:40:51 crc kubenswrapper[4990]: E1203 12:40:51.836222 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826bd110-77e7-484c-bd48-d95fa1b4a0cd" containerName="collect-profiles" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.836241 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="826bd110-77e7-484c-bd48-d95fa1b4a0cd" containerName="collect-profiles" Dec 03 12:40:51 crc kubenswrapper[4990]: E1203 12:40:51.836273 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feff5a1f-1971-4743-aba5-46ae4c2aca8a" containerName="pruner" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.836282 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="feff5a1f-1971-4743-aba5-46ae4c2aca8a" containerName="pruner" Dec 03 12:40:51 crc kubenswrapper[4990]: E1203 12:40:51.836313 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac5c2af-97e9-4e14-8489-4d87435d5856" containerName="pruner" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.836322 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac5c2af-97e9-4e14-8489-4d87435d5856" containerName="pruner" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.836503 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="826bd110-77e7-484c-bd48-d95fa1b4a0cd" containerName="collect-profiles" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.836537 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="feff5a1f-1971-4743-aba5-46ae4c2aca8a" containerName="pruner" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.836550 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac5c2af-97e9-4e14-8489-4d87435d5856" containerName="pruner" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.837067 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.839687 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.840568 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.843111 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.864120 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3060d97f-fdfb-4123-9063-d7e227b4280f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.864198 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3060d97f-fdfb-4123-9063-d7e227b4280f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.973085 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3060d97f-fdfb-4123-9063-d7e227b4280f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.973158 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3060d97f-fdfb-4123-9063-d7e227b4280f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.973334 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3060d97f-fdfb-4123-9063-d7e227b4280f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:51 crc kubenswrapper[4990]: I1203 12:40:51.991437 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3060d97f-fdfb-4123-9063-d7e227b4280f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:52 crc kubenswrapper[4990]: I1203 12:40:52.165516 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:40:54 crc kubenswrapper[4990]: I1203 12:40:54.930335 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:40:54 crc kubenswrapper[4990]: I1203 12:40:54.930712 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.215339 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.216410 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.229255 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.318514 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.318698 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-var-lock\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.318730 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1615d35b-f818-40ae-87c2-59fbb9b73fae-kube-api-access\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.419834 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.420134 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-var-lock\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.420221 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1615d35b-f818-40ae-87c2-59fbb9b73fae-kube-api-access\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.420009 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.420169 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-var-lock\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.439207 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1615d35b-f818-40ae-87c2-59fbb9b73fae-kube-api-access\") pod \"installer-9-crc\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:40:56 crc kubenswrapper[4990]: I1203 12:40:56.543644 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:41:03 crc kubenswrapper[4990]: I1203 12:41:03.286917 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:41:03 crc kubenswrapper[4990]: I1203 12:41:03.287586 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:41:03 crc kubenswrapper[4990]: I1203 12:41:03.287663 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:41:03 crc kubenswrapper[4990]: I1203 12:41:03.289051 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 12:41:03 crc kubenswrapper[4990]: I1203 12:41:03.289201 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156" gracePeriod=600 Dec 03 12:41:04 crc kubenswrapper[4990]: I1203 12:41:04.930060 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:04 crc kubenswrapper[4990]: I1203 12:41:04.930122 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:09 crc kubenswrapper[4990]: I1203 12:41:09.341934 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156" exitCode=0 Dec 03 12:41:09 crc kubenswrapper[4990]: I1203 12:41:09.342033 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156"} Dec 03 12:41:10 crc kubenswrapper[4990]: E1203 12:41:10.295687 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 12:41:10 crc kubenswrapper[4990]: E1203 12:41:10.296239 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8jnqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jxn9m_openshift-marketplace(8108de1d-75ea-419d-9658-e5a035088b36): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:10 crc kubenswrapper[4990]: E1203 12:41:10.297505 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jxn9m" podUID="8108de1d-75ea-419d-9658-e5a035088b36" Dec 03 12:41:10 crc kubenswrapper[4990]: E1203 12:41:10.343985 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 12:41:10 crc kubenswrapper[4990]: E1203 12:41:10.344122 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cc8t9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tprnr_openshift-marketplace(ed381166-1062-4ca5-a553-81ab9681698e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:10 crc kubenswrapper[4990]: E1203 12:41:10.346066 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tprnr" podUID="ed381166-1062-4ca5-a553-81ab9681698e" Dec 03 12:41:12 crc kubenswrapper[4990]: E1203 12:41:12.033700 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jxn9m" podUID="8108de1d-75ea-419d-9658-e5a035088b36" Dec 03 12:41:12 crc kubenswrapper[4990]: E1203 12:41:12.034544 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tprnr" podUID="ed381166-1062-4ca5-a553-81ab9681698e" Dec 03 12:41:12 crc kubenswrapper[4990]: E1203 12:41:12.183077 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 12:41:12 crc kubenswrapper[4990]: E1203 12:41:12.183245 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-466jz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-j98rl_openshift-marketplace(89cb569a-c36f-481c-b5a0-00694046c00c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:12 crc kubenswrapper[4990]: E1203 12:41:12.184526 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-j98rl" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" Dec 03 12:41:13 crc kubenswrapper[4990]: E1203 12:41:13.470976 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-j98rl" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" Dec 03 12:41:13 crc kubenswrapper[4990]: E1203 12:41:13.532343 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 12:41:13 crc kubenswrapper[4990]: E1203 12:41:13.532567 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jd42h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-cbxg8_openshift-marketplace(33b6b642-d3d4-4a33-9bf0-25d5c2aa9030): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:13 crc kubenswrapper[4990]: E1203 12:41:13.533947 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-cbxg8" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" Dec 03 12:41:14 crc kubenswrapper[4990]: I1203 12:41:14.929536 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:14 crc kubenswrapper[4990]: I1203 12:41:14.929900 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:17 crc kubenswrapper[4990]: E1203 12:41:17.818336 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-cbxg8" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.115649 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.116362 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tthkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-sfvnj_openshift-marketplace(2069abff-1d0a-42a0-8929-75acceecbe8b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.117830 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-sfvnj" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.148873 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.149031 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zj55,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dbqqz_openshift-marketplace(ddec4b1a-b34e-45f7-95e6-a04d29dde82c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.152686 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dbqqz" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" Dec 03 12:41:18 crc kubenswrapper[4990]: I1203 12:41:18.255049 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 12:41:18 crc kubenswrapper[4990]: W1203 12:41:18.269870 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3060d97f_fdfb_4123_9063_d7e227b4280f.slice/crio-35ae02ccfd82b2878b759cc33ff2056ebc5d619c6e9a0e1492a01e5cd9bb7ef4 WatchSource:0}: Error finding container 35ae02ccfd82b2878b759cc33ff2056ebc5d619c6e9a0e1492a01e5cd9bb7ef4: Status 404 returned error can't find the container with id 35ae02ccfd82b2878b759cc33ff2056ebc5d619c6e9a0e1492a01e5cd9bb7ef4 Dec 03 12:41:18 crc kubenswrapper[4990]: I1203 12:41:18.396288 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 12:41:18 crc kubenswrapper[4990]: I1203 12:41:18.404359 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3060d97f-fdfb-4123-9063-d7e227b4280f","Type":"ContainerStarted","Data":"35ae02ccfd82b2878b759cc33ff2056ebc5d619c6e9a0e1492a01e5cd9bb7ef4"} Dec 03 12:41:18 crc kubenswrapper[4990]: I1203 12:41:18.411756 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lh8cf" event={"ID":"80543db8-34e7-48cb-93db-897ae58288d2","Type":"ContainerStarted","Data":"cf70b71c0cc1f99de99ddd6eef98bcadadd7cbdc748e6f02b3a2b5ebbdecc318"} Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.412357 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dbqqz" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.413472 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-sfvnj" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.673938 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.674284 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v5pqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kjsgh_openshift-marketplace(572c65be-7211-47d9-805e-081d07fbe844): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.675579 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kjsgh" podUID="572c65be-7211-47d9-805e-081d07fbe844" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.760690 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.761348 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5nrdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zn87g_openshift-marketplace(34bb2446-2a47-4b35-b5a1-13d8de74822e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 12:41:18 crc kubenswrapper[4990]: E1203 12:41:18.762626 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zn87g" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.419839 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"8ff22e0abfe7a9119c84cea1563bea69a1449b575844b4f78fdcfa08778a9149"} Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.421329 4990 generic.go:334] "Generic (PLEG): container finished" podID="3060d97f-fdfb-4123-9063-d7e227b4280f" containerID="4df2b54dde7844b5b2ce5fec60bdce8a7847f424fd21f413d5c2d4d98d6a01ea" exitCode=0 Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.421410 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3060d97f-fdfb-4123-9063-d7e227b4280f","Type":"ContainerDied","Data":"4df2b54dde7844b5b2ce5fec60bdce8a7847f424fd21f413d5c2d4d98d6a01ea"} Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.423951 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1615d35b-f818-40ae-87c2-59fbb9b73fae","Type":"ContainerStarted","Data":"76d6020fc02fd27b128dedbd1052369e9ef703f606d40f981d55d716293ed0ee"} Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.423991 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1615d35b-f818-40ae-87c2-59fbb9b73fae","Type":"ContainerStarted","Data":"0d76e54dd992c49c8bad76f703535847c0e14c90fa55fd11517d8be5c4ed6800"} Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.424866 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.424904 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.425208 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:41:19 crc kubenswrapper[4990]: E1203 12:41:19.427137 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kjsgh" podUID="572c65be-7211-47d9-805e-081d07fbe844" Dec 03 12:41:19 crc kubenswrapper[4990]: E1203 12:41:19.427166 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zn87g" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" Dec 03 12:41:19 crc kubenswrapper[4990]: I1203 12:41:19.528514 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=23.528496072 podStartE2EDuration="23.528496072s" podCreationTimestamp="2025-12-03 12:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:41:19.525190905 +0000 UTC m=+227.667102154" watchObservedRunningTime="2025-12-03 12:41:19.528496072 +0000 UTC m=+227.670407301" Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.431667 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.432527 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.717027 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.753773 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3060d97f-fdfb-4123-9063-d7e227b4280f-kube-api-access\") pod \"3060d97f-fdfb-4123-9063-d7e227b4280f\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.753889 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3060d97f-fdfb-4123-9063-d7e227b4280f-kubelet-dir\") pod \"3060d97f-fdfb-4123-9063-d7e227b4280f\" (UID: \"3060d97f-fdfb-4123-9063-d7e227b4280f\") " Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.754254 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3060d97f-fdfb-4123-9063-d7e227b4280f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3060d97f-fdfb-4123-9063-d7e227b4280f" (UID: "3060d97f-fdfb-4123-9063-d7e227b4280f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.762568 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3060d97f-fdfb-4123-9063-d7e227b4280f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3060d97f-fdfb-4123-9063-d7e227b4280f" (UID: "3060d97f-fdfb-4123-9063-d7e227b4280f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.855298 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3060d97f-fdfb-4123-9063-d7e227b4280f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:41:20 crc kubenswrapper[4990]: I1203 12:41:20.855345 4990 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3060d97f-fdfb-4123-9063-d7e227b4280f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:41:21 crc kubenswrapper[4990]: I1203 12:41:21.435530 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3060d97f-fdfb-4123-9063-d7e227b4280f","Type":"ContainerDied","Data":"35ae02ccfd82b2878b759cc33ff2056ebc5d619c6e9a0e1492a01e5cd9bb7ef4"} Dec 03 12:41:21 crc kubenswrapper[4990]: I1203 12:41:21.435575 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35ae02ccfd82b2878b759cc33ff2056ebc5d619c6e9a0e1492a01e5cd9bb7ef4" Dec 03 12:41:21 crc kubenswrapper[4990]: I1203 12:41:21.435638 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 12:41:24 crc kubenswrapper[4990]: I1203 12:41:24.929682 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:24 crc kubenswrapper[4990]: I1203 12:41:24.929695 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:24 crc kubenswrapper[4990]: I1203 12:41:24.930414 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:24 crc kubenswrapper[4990]: I1203 12:41:24.930544 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:25 crc kubenswrapper[4990]: I1203 12:41:25.904670 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dqrqw"] Dec 03 12:41:34 crc kubenswrapper[4990]: I1203 12:41:34.928836 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:34 crc kubenswrapper[4990]: I1203 12:41:34.929562 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:34 crc kubenswrapper[4990]: I1203 12:41:34.930141 4990 patch_prober.go:28] interesting pod/downloads-7954f5f757-lh8cf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 12:41:34 crc kubenswrapper[4990]: I1203 12:41:34.930165 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lh8cf" podUID="80543db8-34e7-48cb-93db-897ae58288d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 12:41:44 crc kubenswrapper[4990]: I1203 12:41:44.942158 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-lh8cf" Dec 03 12:41:50 crc kubenswrapper[4990]: I1203 12:41:50.661652 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerStarted","Data":"c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589"} Dec 03 12:41:50 crc kubenswrapper[4990]: I1203 12:41:50.663445 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerStarted","Data":"daf286fe6684453d7969af2d97d03333d4c9d9106f671079d77dd8c342eae668"} Dec 03 12:41:50 crc kubenswrapper[4990]: I1203 12:41:50.665414 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerStarted","Data":"4542dd6aa8c3f3b41dc07b034ae5e13683ffb8df40a97d28633fcee6e857b984"} Dec 03 12:41:50 crc kubenswrapper[4990]: I1203 12:41:50.667079 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerStarted","Data":"2a1485703ee2e7f0106e40143e7b7187d8344aa0fa2637a341d46dceccb44993"} Dec 03 12:41:50 crc kubenswrapper[4990]: I1203 12:41:50.952955 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" containerID="cri-o://cc935d040ebc2ca4be0665585ec1a53eb10a7f990431c939655aca43286b536b" gracePeriod=15 Dec 03 12:41:51 crc kubenswrapper[4990]: I1203 12:41:51.674397 4990 generic.go:334] "Generic (PLEG): container finished" podID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerID="cc935d040ebc2ca4be0665585ec1a53eb10a7f990431c939655aca43286b536b" exitCode=0 Dec 03 12:41:51 crc kubenswrapper[4990]: I1203 12:41:51.674492 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" event={"ID":"2ccf32a0-8701-4d6f-8c45-d314f98754b3","Type":"ContainerDied","Data":"cc935d040ebc2ca4be0665585ec1a53eb10a7f990431c939655aca43286b536b"} Dec 03 12:41:53 crc kubenswrapper[4990]: I1203 12:41:53.686749 4990 generic.go:334] "Generic (PLEG): container finished" podID="ed381166-1062-4ca5-a553-81ab9681698e" containerID="c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589" exitCode=0 Dec 03 12:41:53 crc kubenswrapper[4990]: I1203 12:41:53.686818 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerDied","Data":"c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589"} Dec 03 12:41:53 crc kubenswrapper[4990]: I1203 12:41:53.689590 4990 generic.go:334] "Generic (PLEG): container finished" podID="8108de1d-75ea-419d-9658-e5a035088b36" containerID="daf286fe6684453d7969af2d97d03333d4c9d9106f671079d77dd8c342eae668" exitCode=0 Dec 03 12:41:53 crc kubenswrapper[4990]: I1203 12:41:53.689642 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerDied","Data":"daf286fe6684453d7969af2d97d03333d4c9d9106f671079d77dd8c342eae668"} Dec 03 12:41:53 crc kubenswrapper[4990]: I1203 12:41:53.711077 4990 generic.go:334] "Generic (PLEG): container finished" podID="89cb569a-c36f-481c-b5a0-00694046c00c" containerID="2a1485703ee2e7f0106e40143e7b7187d8344aa0fa2637a341d46dceccb44993" exitCode=0 Dec 03 12:41:53 crc kubenswrapper[4990]: I1203 12:41:53.711153 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerDied","Data":"2a1485703ee2e7f0106e40143e7b7187d8344aa0fa2637a341d46dceccb44993"} Dec 03 12:41:55 crc kubenswrapper[4990]: I1203 12:41:55.118809 4990 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dqrqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 03 12:41:55 crc kubenswrapper[4990]: I1203 12:41:55.118870 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.611474 4990 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.612048 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3060d97f-fdfb-4123-9063-d7e227b4280f" containerName="pruner" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.612065 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="3060d97f-fdfb-4123-9063-d7e227b4280f" containerName="pruner" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.612194 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="3060d97f-fdfb-4123-9063-d7e227b4280f" containerName="pruner" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.612650 4990 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.612757 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613050 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276" gracePeriod=15 Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613139 4990 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613172 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5" gracePeriod=15 Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613281 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645" gracePeriod=15 Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613338 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2" gracePeriod=15 Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613185 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018" gracePeriod=15 Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613344 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613476 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613487 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613494 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613505 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613511 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613519 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613525 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613538 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613544 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613554 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613560 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 12:41:56 crc kubenswrapper[4990]: E1203 12:41:56.613568 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.613761 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.614217 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.614238 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.614251 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.614264 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.614277 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.614536 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.616386 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.616672 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.616747 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.616885 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.616935 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.657193 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.717829 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.717897 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.717924 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.717975 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718004 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718042 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718051 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718085 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718063 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718121 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718153 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718163 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.718167 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.818694 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.818757 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.818829 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.818850 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.818874 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.818929 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:41:56 crc kubenswrapper[4990]: I1203 12:41:56.951433 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.736686 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.738219 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.739049 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018" exitCode=2 Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.742003 4990 generic.go:334] "Generic (PLEG): container finished" podID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerID="4542dd6aa8c3f3b41dc07b034ae5e13683ffb8df40a97d28633fcee6e857b984" exitCode=0 Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.742090 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerDied","Data":"4542dd6aa8c3f3b41dc07b034ae5e13683ffb8df40a97d28633fcee6e857b984"} Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.743296 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.743910 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:41:57 crc kubenswrapper[4990]: I1203 12:41:57.744284 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:41:58 crc kubenswrapper[4990]: I1203 12:41:58.749625 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 12:41:58 crc kubenswrapper[4990]: I1203 12:41:58.751229 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 12:41:58 crc kubenswrapper[4990]: I1203 12:41:58.751919 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645" exitCode=0 Dec 03 12:41:58 crc kubenswrapper[4990]: I1203 12:41:58.751973 4990 scope.go:117] "RemoveContainer" containerID="62fa6757537b14f7bb4583de93562dc575b02f4b72bd68a7120fca2835451d80" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.319938 4990 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.320423 4990 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.320653 4990 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.320884 4990 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.321083 4990 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:00 crc kubenswrapper[4990]: I1203 12:42:00.321115 4990 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.321344 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="200ms" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.522904 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="400ms" Dec 03 12:42:00 crc kubenswrapper[4990]: E1203 12:42:00.924033 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="800ms" Dec 03 12:42:01 crc kubenswrapper[4990]: E1203 12:42:01.725344 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="1.6s" Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.776522 4990 generic.go:334] "Generic (PLEG): container finished" podID="1615d35b-f818-40ae-87c2-59fbb9b73fae" containerID="76d6020fc02fd27b128dedbd1052369e9ef703f606d40f981d55d716293ed0ee" exitCode=0 Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.776611 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1615d35b-f818-40ae-87c2-59fbb9b73fae","Type":"ContainerDied","Data":"76d6020fc02fd27b128dedbd1052369e9ef703f606d40f981d55d716293ed0ee"} Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.777477 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.778295 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.778631 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.780253 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.781389 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2" exitCode=0 Dec 03 12:42:01 crc kubenswrapper[4990]: I1203 12:42:01.781442 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5" exitCode=0 Dec 03 12:42:02 crc kubenswrapper[4990]: I1203 12:42:02.267695 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:02 crc kubenswrapper[4990]: I1203 12:42:02.268704 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:02 crc kubenswrapper[4990]: I1203 12:42:02.269478 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:03 crc kubenswrapper[4990]: E1203 12:42:03.325911 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="3.2s" Dec 03 12:42:03 crc kubenswrapper[4990]: I1203 12:42:03.793920 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 12:42:03 crc kubenswrapper[4990]: I1203 12:42:03.794666 4990 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276" exitCode=0 Dec 03 12:42:05 crc kubenswrapper[4990]: I1203 12:42:05.118736 4990 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dqrqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 03 12:42:05 crc kubenswrapper[4990]: I1203 12:42:05.119304 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 03 12:42:05 crc kubenswrapper[4990]: E1203 12:42:05.120074 4990 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events/oauth-openshift-558db77b4-dqrqw.187db4f823a2cda0\": dial tcp 38.102.83.130:6443: connect: connection refused" event=< Dec 03 12:42:05 crc kubenswrapper[4990]: &Event{ObjectMeta:{oauth-openshift-558db77b4-dqrqw.187db4f823a2cda0 openshift-authentication 29278 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-558db77b4-dqrqw,UID:2ccf32a0-8701-4d6f-8c45-d314f98754b3,APIVersion:v1,ResourceVersion:27206,FieldPath:spec.containers{oauth-openshift},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.8:6443/healthz": dial tcp 10.217.0.8:6443: connect: connection refused Dec 03 12:42:05 crc kubenswrapper[4990]: body: Dec 03 12:42:05 crc kubenswrapper[4990]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 12:40:06 +0000 UTC,LastTimestamp:2025-12-03 12:42:05.119268302 +0000 UTC m=+273.261179551,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 12:42:05 crc kubenswrapper[4990]: > Dec 03 12:42:05 crc kubenswrapper[4990]: E1203 12:42:05.334625 4990 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.130:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" volumeName="registry-storage" Dec 03 12:42:06 crc kubenswrapper[4990]: E1203 12:42:06.527700 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="6.4s" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.410478 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.411441 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.411974 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.412241 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.479424 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.480736 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.481625 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.481844 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.482183 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.482723 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.564957 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1615d35b-f818-40ae-87c2-59fbb9b73fae-kube-api-access\") pod \"1615d35b-f818-40ae-87c2-59fbb9b73fae\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565338 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565365 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565491 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565522 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565541 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-var-lock\") pod \"1615d35b-f818-40ae-87c2-59fbb9b73fae\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565578 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-var-lock" (OuterVolumeSpecName: "var-lock") pod "1615d35b-f818-40ae-87c2-59fbb9b73fae" (UID: "1615d35b-f818-40ae-87c2-59fbb9b73fae"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565586 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565651 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565697 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1615d35b-f818-40ae-87c2-59fbb9b73fae" (UID: "1615d35b-f818-40ae-87c2-59fbb9b73fae"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.565679 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-kubelet-dir\") pod \"1615d35b-f818-40ae-87c2-59fbb9b73fae\" (UID: \"1615d35b-f818-40ae-87c2-59fbb9b73fae\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.566284 4990 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.566300 4990 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.566309 4990 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.566318 4990 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1615d35b-f818-40ae-87c2-59fbb9b73fae-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.566327 4990 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.573387 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1615d35b-f818-40ae-87c2-59fbb9b73fae-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1615d35b-f818-40ae-87c2-59fbb9b73fae" (UID: "1615d35b-f818-40ae-87c2-59fbb9b73fae"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.648883 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.649397 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.649799 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.650612 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.650913 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.651315 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: W1203 12:42:07.653234 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-dafec3d33523464d89a1c44e0ef930c3f835b47e16fbfc7bb6fa1c84799beddc WatchSource:0}: Error finding container dafec3d33523464d89a1c44e0ef930c3f835b47e16fbfc7bb6fa1c84799beddc: Status 404 returned error can't find the container with id dafec3d33523464d89a1c44e0ef930c3f835b47e16fbfc7bb6fa1c84799beddc Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667285 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-cliconfig\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667347 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-policies\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667383 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-idp-0-file-data\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667412 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-trusted-ca-bundle\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667437 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-session\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667488 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-serving-cert\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667534 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-router-certs\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667605 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-login\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667650 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-dir\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667676 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttwfl\" (UniqueName: \"kubernetes.io/projected/2ccf32a0-8701-4d6f-8c45-d314f98754b3-kube-api-access-ttwfl\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667706 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-error\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667733 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-service-ca\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667766 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-ocp-branding-template\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667794 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-provider-selection\") pod \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\" (UID: \"2ccf32a0-8701-4d6f-8c45-d314f98754b3\") " Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.667997 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1615d35b-f818-40ae-87c2-59fbb9b73fae-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.668572 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.669217 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.669269 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.669956 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.670916 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.671252 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.671620 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.672146 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ccf32a0-8701-4d6f-8c45-d314f98754b3-kube-api-access-ttwfl" (OuterVolumeSpecName: "kube-api-access-ttwfl") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "kube-api-access-ttwfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.672983 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.673713 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.674622 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.674723 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.676878 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.678370 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2ccf32a0-8701-4d6f-8c45-d314f98754b3" (UID: "2ccf32a0-8701-4d6f-8c45-d314f98754b3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.769351 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.769417 4990 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.769433 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770068 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770169 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770185 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770199 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770214 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770226 4990 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2ccf32a0-8701-4d6f-8c45-d314f98754b3-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770240 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttwfl\" (UniqueName: \"kubernetes.io/projected/2ccf32a0-8701-4d6f-8c45-d314f98754b3-kube-api-access-ttwfl\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770255 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770269 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770287 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.770300 4990 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2ccf32a0-8701-4d6f-8c45-d314f98754b3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.827475 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.829066 4990 scope.go:117] "RemoveContainer" containerID="7b98708e885d2596ee8791ba53661204b607463c1ca71ca3a30556d04e64a645" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.829086 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.830249 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.830834 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.831078 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.831327 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.831615 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.834588 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerStarted","Data":"72aad500c91a6505c1c7d762fc1a45efaaa852347212d9c2cf391a8ae7a7b09f"} Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.836393 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.837066 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.837375 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.838428 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.838980 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.839206 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.840262 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" event={"ID":"2ccf32a0-8701-4d6f-8c45-d314f98754b3","Type":"ContainerDied","Data":"c7f142e6167dfaedad8b9a618b442c8e54615e64d81e617f286b1362313d2caa"} Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.840359 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.842572 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.844802 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.845564 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.845850 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.846083 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.846589 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.847229 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.847430 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.847712 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1615d35b-f818-40ae-87c2-59fbb9b73fae","Type":"ContainerDied","Data":"0d76e54dd992c49c8bad76f703535847c0e14c90fa55fd11517d8be5c4ed6800"} Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.847753 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d76e54dd992c49c8bad76f703535847c0e14c90fa55fd11517d8be5c4ed6800" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.847746 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.847879 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.848024 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.848217 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.848412 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.850860 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerStarted","Data":"9be9f927b0eb3489b7566fcd0bdb56ad3c9ea2b8bd01ffbcb38959c5d63ba73b"} Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.852150 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"dafec3d33523464d89a1c44e0ef930c3f835b47e16fbfc7bb6fa1c84799beddc"} Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.858311 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.858900 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.859531 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.859766 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.860204 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.861179 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.872126 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.873322 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.874779 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.875179 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.875580 4990 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.875873 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:07 crc kubenswrapper[4990]: I1203 12:42:07.954742 4990 scope.go:117] "RemoveContainer" containerID="13f0b860260efa7bf1e31cb7b1f39a8fc48b02d7e001ff86e4bfbd7ac4ebf5b2" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.109355 4990 scope.go:117] "RemoveContainer" containerID="9572b2471168db51dccf410b0078eb07270023407538515e8d65bbb991fe22b5" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.137942 4990 scope.go:117] "RemoveContainer" containerID="bcf4c9ce3711b43d052399acbd6c7ac6a67d194e03a70a23f6cc1f7a86bb3018" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.271957 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.441575 4990 scope.go:117] "RemoveContainer" containerID="4d322a761397bf608c562c65943582e4155ad77afb3561189d3f50f7a7faa276" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.500062 4990 scope.go:117] "RemoveContainer" containerID="3f19606d89acafbd6db5267e0bd1478548a60813f2b32eb0b3b9583bac5fb222" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.566754 4990 scope.go:117] "RemoveContainer" containerID="cc935d040ebc2ca4be0665585ec1a53eb10a7f990431c939655aca43286b536b" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.861677 4990 generic.go:334] "Generic (PLEG): container finished" podID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerID="8d1cddaff8fc11f602e6bba7598aa77a3eb2bb33e5d5f233160020c224883073" exitCode=0 Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.861798 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbxg8" event={"ID":"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030","Type":"ContainerDied","Data":"8d1cddaff8fc11f602e6bba7598aa77a3eb2bb33e5d5f233160020c224883073"} Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.863508 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.864120 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.864809 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.865125 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.865410 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.865809 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.867015 4990 generic.go:334] "Generic (PLEG): container finished" podID="572c65be-7211-47d9-805e-081d07fbe844" containerID="9be9f927b0eb3489b7566fcd0bdb56ad3c9ea2b8bd01ffbcb38959c5d63ba73b" exitCode=0 Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.867137 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerDied","Data":"9be9f927b0eb3489b7566fcd0bdb56ad3c9ea2b8bd01ffbcb38959c5d63ba73b"} Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.867797 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.868116 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.868761 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.869157 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.869386 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"aff99b436d50dc5c6e07784235b5da37c236aea6f8afacc7939bb83e42234f9e"} Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.869618 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.869994 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.870394 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.871072 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.871550 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.871993 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.872347 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.872602 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.872814 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.872855 4990 generic.go:334] "Generic (PLEG): container finished" podID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerID="2b9b4f697a3b4a26f7dd6745d64d07970e5e1e5040d084b650145ac110a347d8" exitCode=0 Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.872902 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbqqz" event={"ID":"ddec4b1a-b34e-45f7-95e6-a04d29dde82c","Type":"ContainerDied","Data":"2b9b4f697a3b4a26f7dd6745d64d07970e5e1e5040d084b650145ac110a347d8"} Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.873022 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.873218 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.873363 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.873705 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.873973 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.874263 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.874716 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.875037 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:08 crc kubenswrapper[4990]: I1203 12:42:08.875264 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.881919 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerStarted","Data":"02d693cb3abf43b38366de04886fea8d8fcf8aa59103b1ac5530b5cd332785e6"} Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.883052 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.883333 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.883653 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.883977 4990 generic.go:334] "Generic (PLEG): container finished" podID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerID="72aad500c91a6505c1c7d762fc1a45efaaa852347212d9c2cf391a8ae7a7b09f" exitCode=0 Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.884095 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerDied","Data":"72aad500c91a6505c1c7d762fc1a45efaaa852347212d9c2cf391a8ae7a7b09f"} Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.884182 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.884509 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.884934 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.885566 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.886112 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.886379 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.886434 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerStarted","Data":"a594425cfa5eb664e8cb7d9236624bfd80eb31ec1f7c0c254b9e9902a0fc09e3"} Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.886759 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.887005 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.887268 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.887571 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.887797 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.887983 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.888208 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.888404 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.888610 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.888716 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerStarted","Data":"8f42b33dc07af86a83a56129107ca063742b5b604584ab5cfb4584c16ef8fc3a"} Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.888938 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.889139 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.889314 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.892554 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.892810 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.893006 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.893211 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.893514 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.893835 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.894223 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.894445 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.894654 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.894841 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.894897 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerStarted","Data":"ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe"} Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.895027 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.895205 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.895414 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.895633 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.895826 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.895975 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.896162 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.896305 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.896468 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.897580 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.897872 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.898375 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.898816 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.899040 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.899335 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.899662 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:09 crc kubenswrapper[4990]: I1203 12:42:09.899875 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:10 crc kubenswrapper[4990]: I1203 12:42:10.623985 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:42:10 crc kubenswrapper[4990]: I1203 12:42:10.624063 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.264363 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.265685 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.266216 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.266395 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.266568 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.266884 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.267391 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.267777 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.267966 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.268172 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.268344 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.268580 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.277916 4990 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.277955 4990 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:11 crc kubenswrapper[4990]: E1203 12:42:11.278795 4990 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:11 crc kubenswrapper[4990]: I1203 12:42:11.279337 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:11 crc kubenswrapper[4990]: E1203 12:42:11.293887 4990 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events/oauth-openshift-558db77b4-dqrqw.187db4f823a2cda0\": dial tcp 38.102.83.130:6443: connect: connection refused" event=< Dec 03 12:42:11 crc kubenswrapper[4990]: &Event{ObjectMeta:{oauth-openshift-558db77b4-dqrqw.187db4f823a2cda0 openshift-authentication 29278 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-558db77b4-dqrqw,UID:2ccf32a0-8701-4d6f-8c45-d314f98754b3,APIVersion:v1,ResourceVersion:27206,FieldPath:spec.containers{oauth-openshift},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.8:6443/healthz": dial tcp 10.217.0.8:6443: connect: connection refused Dec 03 12:42:11 crc kubenswrapper[4990]: body: Dec 03 12:42:11 crc kubenswrapper[4990]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 12:40:06 +0000 UTC,LastTimestamp:2025-12-03 12:42:05.119268302 +0000 UTC m=+273.261179551,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 12:42:11 crc kubenswrapper[4990]: > Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.271531 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.271970 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.272394 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.272941 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.273187 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.273560 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.274030 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.274344 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.274649 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.275515 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.275810 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:12 crc kubenswrapper[4990]: I1203 12:42:12.276094 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.925773 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.926017 4990 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1" exitCode=1 Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.926047 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1"} Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.926594 4990 scope.go:117] "RemoveContainer" containerID="9c903a0c473e843afea3b64f75a031c2c838d8af872b929fd6f8c64b17fc9bb1" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.926998 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.927500 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.927899 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.928158 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.928365 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:12.928697 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="7s" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.928741 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.929098 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.929397 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.929740 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.930047 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.930311 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.931103 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:12.931662 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:13.340116 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sfvnj" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="registry-server" probeResult="failure" output=< Dec 03 12:42:25 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 12:42:25 crc kubenswrapper[4990]: > Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:13.433022 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:15.737775 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:16.988235 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:16.988276 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.182928 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.183819 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.184169 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.184694 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.185014 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.185297 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.185654 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.186151 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.186361 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.186626 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.186980 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.187337 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.187636 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.187898 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.569584 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.569655 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.613271 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.614015 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.614663 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.615189 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.615682 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.616009 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.616349 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.616656 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.616945 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.617291 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.617614 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.617952 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.618234 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.618515 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.995194 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.995244 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.995928 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.996360 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.996628 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.996820 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.996996 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.997175 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.997394 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.997818 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.998267 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.998569 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.999461 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.999685 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:17.999929 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.000415 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.000698 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.000968 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.001258 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.001639 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.001908 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.002183 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.002499 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.002759 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.003051 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.003293 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.003564 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.003836 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.980628 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:18.980663 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.015815 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.016376 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.016835 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.017203 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.017793 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.018405 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.018796 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.019145 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.019616 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.020007 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.020279 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.020679 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.021070 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:19.021378 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:19.929568 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="7s" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.004646 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.005298 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.005704 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.006315 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.007926 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.008817 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.009890 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.010837 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.011875 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.013058 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.023939 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.025990 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.026690 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.027300 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.664501 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.665512 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.665955 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.666278 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.666614 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.666857 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.667172 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.667503 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.667752 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.668035 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.668310 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.668566 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.668781 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.669012 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.710490 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.711243 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.711662 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.711956 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.712371 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.712763 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.713052 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.713344 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.713598 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.713953 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.714794 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.715117 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.715424 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:20.715700 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:21.239360 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:21.294647 4990 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events/oauth-openshift-558db77b4-dqrqw.187db4f823a2cda0\": dial tcp 38.102.83.130:6443: connect: connection refused" event=< Dec 03 12:42:25 crc kubenswrapper[4990]: &Event{ObjectMeta:{oauth-openshift-558db77b4-dqrqw.187db4f823a2cda0 openshift-authentication 29278 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-558db77b4-dqrqw,UID:2ccf32a0-8701-4d6f-8c45-d314f98754b3,APIVersion:v1,ResourceVersion:27206,FieldPath:spec.containers{oauth-openshift},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.8:6443/healthz": dial tcp 10.217.0.8:6443: connect: connection refused Dec 03 12:42:25 crc kubenswrapper[4990]: body: Dec 03 12:42:25 crc kubenswrapper[4990]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 12:40:06 +0000 UTC,LastTimestamp:2025-12-03 12:42:05.119268302 +0000 UTC m=+273.261179551,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 12:42:25 crc kubenswrapper[4990]: > Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.266667 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.267168 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.267861 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.268315 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.268826 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.269156 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.269679 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.270126 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.270588 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.271039 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.271639 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.272077 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: I1203 12:42:22.272377 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:25.525329 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:42:25Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:42:25Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:42:25Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T12:42:25Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[],\\\"sizeBytes\\\":1609873225},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:446e5d504e70c7963ef7b0f090f3fcb19847ef48150299bf030847565d7a579b\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a01ee07f4838bab6cfa5a3d25d867557aa271725bfcd20a1e52d3cc63423c06b\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1204969293},{\\\"names\\\":[],\\\"sizeBytes\\\":1201319250},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:25.526165 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:25.526794 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:25.527188 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:25.527545 4990 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:25 crc kubenswrapper[4990]: E1203 12:42:25.527572 4990 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 12:42:26 crc kubenswrapper[4990]: E1203 12:42:26.931043 4990 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.130:6443: connect: connection refused" interval="7s" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.011794 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerStarted","Data":"abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.013397 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.013858 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.014293 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.015194 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.015423 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.015647 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.015848 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.017253 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.017496 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.017787 4990 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="46c1bc81e21fa2b50631cd4b68734c1edf76e0a21e9559900948aa725dba3290" exitCode=0 Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.017820 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"46c1bc81e21fa2b50631cd4b68734c1edf76e0a21e9559900948aa725dba3290"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.017853 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f340f5c1c0cb4d5a00b8dd5e5c48431e10e5119dc37356e6a07b54fdf398c7b4"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.017788 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.018157 4990 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.018192 4990 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.018353 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: E1203 12:42:27.018577 4990 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.018850 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.019066 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.019422 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.019755 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.020014 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.020213 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.020393 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.020606 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.020793 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.020975 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.021173 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.021375 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.021623 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.021945 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.022221 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.025664 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbxg8" event={"ID":"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030","Type":"ContainerStarted","Data":"8d6079685da74fad7c22d1fbec9328b4600943f35c4084c3b70e30aa3cc66ddf"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.026325 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.026640 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.026907 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.027113 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.027284 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.027433 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.027617 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.027767 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.027920 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.028066 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.028236 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.028411 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.028651 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.033770 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerStarted","Data":"b3a06d33410d58cd038e690a65230a7969614cb6299d1357f88d8fe8ff737a0a"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.034615 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.034950 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.035289 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.035540 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.035747 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.035952 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.036165 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.036350 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.036643 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.036900 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.037120 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.037377 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.037697 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.040581 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.040962 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"63180b9122d996ad16057dd87045a83e0c9d438c4a72aed6910a8f465cc414e9"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.041885 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.042158 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.043059 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.043296 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.043525 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.043740 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.043941 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.044159 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.044340 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.044512 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.044648 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.044847 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.045055 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.045393 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbqqz" event={"ID":"ddec4b1a-b34e-45f7-95e6-a04d29dde82c","Type":"ContainerStarted","Data":"98ce943cfb69275a66c986eafff99bc65a3a80144327c59fb8fa0c49f9df9f23"} Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.046086 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.046244 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.046444 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.046773 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.046988 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.047318 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.047749 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.047960 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.048159 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.048362 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.052628 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.053169 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.053560 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.368804 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:42:27 crc kubenswrapper[4990]: I1203 12:42:27.368890 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:42:28 crc kubenswrapper[4990]: I1203 12:42:28.409514 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-kjsgh" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="registry-server" probeResult="failure" output=< Dec 03 12:42:28 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 12:42:28 crc kubenswrapper[4990]: > Dec 03 12:42:28 crc kubenswrapper[4990]: I1203 12:42:28.957986 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:42:28 crc kubenswrapper[4990]: I1203 12:42:28.958056 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.024306 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.025005 4990 status_manager.go:851] "Failed to get status for pod" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" pod="openshift-marketplace/redhat-operators-sfvnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-sfvnj\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.025378 4990 status_manager.go:851] "Failed to get status for pod" podUID="572c65be-7211-47d9-805e-081d07fbe844" pod="openshift-marketplace/community-operators-kjsgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-kjsgh\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.025742 4990 status_manager.go:851] "Failed to get status for pod" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" pod="openshift-marketplace/redhat-marketplace-cbxg8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cbxg8\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.026147 4990 status_manager.go:851] "Failed to get status for pod" podUID="8108de1d-75ea-419d-9658-e5a035088b36" pod="openshift-marketplace/certified-operators-jxn9m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jxn9m\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.026326 4990 status_manager.go:851] "Failed to get status for pod" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" pod="openshift-marketplace/redhat-operators-zn87g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-zn87g\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.026577 4990 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.026765 4990 status_manager.go:851] "Failed to get status for pod" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" pod="openshift-marketplace/community-operators-j98rl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j98rl\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.026938 4990 status_manager.go:851] "Failed to get status for pod" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" pod="openshift-authentication/oauth-openshift-558db77b4-dqrqw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dqrqw\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.027089 4990 status_manager.go:851] "Failed to get status for pod" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.027245 4990 status_manager.go:851] "Failed to get status for pod" podUID="ed381166-1062-4ca5-a553-81ab9681698e" pod="openshift-marketplace/certified-operators-tprnr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-tprnr\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.027391 4990 status_manager.go:851] "Failed to get status for pod" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" pod="openshift-marketplace/redhat-marketplace-dbqqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dbqqz\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.027675 4990 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.027902 4990 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.130:6443: connect: connection refused" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.065965 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b24ff9f65b12ce12d9bcd85f8e6da1b254e4a9c9841c576619d1ff99b9dc0f65"} Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.640338 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.641500 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:42:29 crc kubenswrapper[4990]: I1203 12:42:29.717213 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.077622 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"24218fa57097115195ee8f0eafe50e731dba9465888835502cc212b1b73d7674"} Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.077963 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"46fb99863e141863e5be60adf79714d1003faa2f523a7041deaa3329535b23a0"} Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.078082 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"345b697c85712c62a3f2673cbebd94498c3cc223f3066c17f44f6dc76c2efdd2"} Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.078167 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c215ce6509ed13967aab57d4b859d77a59e579b9c10e12f78ac7e3ba8a1a4899"} Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.078176 4990 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.078344 4990 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.197878 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:42:30 crc kubenswrapper[4990]: I1203 12:42:30.197973 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.245169 4990 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.245236 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.245321 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.266631 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zn87g" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="registry-server" probeResult="failure" output=< Dec 03 12:42:31 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 12:42:31 crc kubenswrapper[4990]: > Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.280063 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.280102 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.280228 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.333991 4990 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.334100 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 12:42:31 crc kubenswrapper[4990]: I1203 12:42:31.377991 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:42:35 crc kubenswrapper[4990]: I1203 12:42:35.737797 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:35 crc kubenswrapper[4990]: I1203 12:42:35.918331 4990 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:42:36 crc kubenswrapper[4990]: I1203 12:42:36.142498 4990 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e1be69eb-5b82-4cee-a850-9a8a4b985c78" Dec 03 12:42:36 crc kubenswrapper[4990]: I1203 12:42:36.307070 4990 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:36 crc kubenswrapper[4990]: I1203 12:42:36.307337 4990 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:42:36 crc kubenswrapper[4990]: I1203 12:42:36.309891 4990 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e1be69eb-5b82-4cee-a850-9a8a4b985c78" Dec 03 12:42:37 crc kubenswrapper[4990]: I1203 12:42:37.422514 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:42:37 crc kubenswrapper[4990]: I1203 12:42:37.463777 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:42:39 crc kubenswrapper[4990]: I1203 12:42:39.022415 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:42:40 crc kubenswrapper[4990]: I1203 12:42:40.240055 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:42:40 crc kubenswrapper[4990]: I1203 12:42:40.279411 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:42:41 crc kubenswrapper[4990]: I1203 12:42:41.243519 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:41 crc kubenswrapper[4990]: I1203 12:42:41.255349 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 12:42:59 crc kubenswrapper[4990]: I1203 12:42:59.318742 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 12:43:00 crc kubenswrapper[4990]: I1203 12:43:00.197038 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 12:43:00 crc kubenswrapper[4990]: I1203 12:43:00.205833 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 12:43:00 crc kubenswrapper[4990]: I1203 12:43:00.699344 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 12:43:01 crc kubenswrapper[4990]: I1203 12:43:01.044138 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 12:43:01 crc kubenswrapper[4990]: I1203 12:43:01.328838 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 12:43:01 crc kubenswrapper[4990]: I1203 12:43:01.399069 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 12:43:02 crc kubenswrapper[4990]: I1203 12:43:02.973156 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 12:43:03 crc kubenswrapper[4990]: I1203 12:43:03.317629 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 12:43:03 crc kubenswrapper[4990]: I1203 12:43:03.663821 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 12:43:03 crc kubenswrapper[4990]: I1203 12:43:03.870801 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 12:43:04 crc kubenswrapper[4990]: I1203 12:43:04.224841 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 12:43:04 crc kubenswrapper[4990]: I1203 12:43:04.326543 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 12:43:04 crc kubenswrapper[4990]: I1203 12:43:04.915199 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 12:43:04 crc kubenswrapper[4990]: I1203 12:43:04.996683 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.110142 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.150047 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.153570 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.409370 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.536888 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.634056 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.688399 4990 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.869818 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 12:43:05 crc kubenswrapper[4990]: I1203 12:43:05.934680 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.019535 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.019671 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.071110 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.077865 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.186702 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.249526 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.462255 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.597760 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.765875 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.861319 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 12:43:06 crc kubenswrapper[4990]: I1203 12:43:06.984710 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 12:43:07 crc kubenswrapper[4990]: I1203 12:43:07.184578 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 12:43:07 crc kubenswrapper[4990]: I1203 12:43:07.429509 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 12:43:07 crc kubenswrapper[4990]: I1203 12:43:07.812401 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 12:43:07 crc kubenswrapper[4990]: I1203 12:43:07.853845 4990 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 12:43:07 crc kubenswrapper[4990]: I1203 12:43:07.854834 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 12:43:07 crc kubenswrapper[4990]: I1203 12:43:07.925880 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.002227 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.119862 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.228394 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.263171 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.331982 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.414564 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.498062 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.806083 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 12:43:08 crc kubenswrapper[4990]: I1203 12:43:08.838126 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.026854 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.205028 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.234039 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.435957 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.603391 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.686502 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.782066 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 12:43:09 crc kubenswrapper[4990]: I1203 12:43:09.859361 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.082485 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.179557 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.191081 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.514853 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.613621 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.663409 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.801352 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.952911 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 12:43:10 crc kubenswrapper[4990]: I1203 12:43:10.960951 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.041534 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.274653 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.359841 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.457419 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.522536 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.750471 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.851709 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 12:43:11 crc kubenswrapper[4990]: I1203 12:43:11.892422 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.018796 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.041780 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.208587 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.214623 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.232817 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.251036 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.281838 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.501161 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.530165 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.604586 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 12:43:12 crc kubenswrapper[4990]: I1203 12:43:12.735128 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.150794 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.211628 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.359367 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.426936 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.746553 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.770433 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.772342 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.831485 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 12:43:13 crc kubenswrapper[4990]: I1203 12:43:13.861399 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.012331 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.026527 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.255229 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.349320 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.383101 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.537496 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.911087 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.965774 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 12:43:14 crc kubenswrapper[4990]: I1203 12:43:14.994528 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.038310 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.250545 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.269208 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.393885 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.501985 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.635749 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 12:43:15 crc kubenswrapper[4990]: I1203 12:43:15.685197 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.033371 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.090146 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.126352 4990 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gmsz9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.126720 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.129143 4990 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gmsz9 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.129318 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.145346 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.259785 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.270676 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.284039 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.288874 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.325191 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.448331 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.511824 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.597732 4990 generic.go:334] "Generic (PLEG): container finished" podID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerID="dad46f22a4c84e3cb5e727b7b026fcc963fb293733d186a1db88d671df532985" exitCode=0 Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.597842 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerDied","Data":"dad46f22a4c84e3cb5e727b7b026fcc963fb293733d186a1db88d671df532985"} Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.598726 4990 scope.go:117] "RemoveContainer" containerID="dad46f22a4c84e3cb5e727b7b026fcc963fb293733d186a1db88d671df532985" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.616950 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.634955 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.684992 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 12:43:16 crc kubenswrapper[4990]: I1203 12:43:16.696904 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.054252 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.075767 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.154840 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.192569 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.211181 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.260199 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.281695 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.319570 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.343385 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.522151 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.595697 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.628756 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gmsz9_2870df3c-124c-4a2d-8f88-4fec5f65740c/marketplace-operator/1.log" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.629342 4990 generic.go:334] "Generic (PLEG): container finished" podID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerID="caadcf267bb9499dd5a413687a3708feda6c22ed87ce1f1f4706f019af7fec70" exitCode=1 Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.629390 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerDied","Data":"caadcf267bb9499dd5a413687a3708feda6c22ed87ce1f1f4706f019af7fec70"} Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.629432 4990 scope.go:117] "RemoveContainer" containerID="dad46f22a4c84e3cb5e727b7b026fcc963fb293733d186a1db88d671df532985" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.630098 4990 scope.go:117] "RemoveContainer" containerID="caadcf267bb9499dd5a413687a3708feda6c22ed87ce1f1f4706f019af7fec70" Dec 03 12:43:17 crc kubenswrapper[4990]: E1203 12:43:17.630333 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-gmsz9_openshift-marketplace(2870df3c-124c-4a2d-8f88-4fec5f65740c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" Dec 03 12:43:17 crc kubenswrapper[4990]: I1203 12:43:17.859506 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.031798 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.063888 4990 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.458563 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.511550 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.589537 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.595600 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.635603 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gmsz9_2870df3c-124c-4a2d-8f88-4fec5f65740c/marketplace-operator/1.log" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.656535 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.694778 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.833564 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.833593 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 12:43:18 crc kubenswrapper[4990]: I1203 12:43:18.969763 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 12:43:19 crc kubenswrapper[4990]: I1203 12:43:19.022155 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 12:43:19 crc kubenswrapper[4990]: I1203 12:43:19.033555 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 12:43:19 crc kubenswrapper[4990]: I1203 12:43:19.106588 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 12:43:19 crc kubenswrapper[4990]: I1203 12:43:19.410066 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 12:43:19 crc kubenswrapper[4990]: I1203 12:43:19.632366 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.010495 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.157214 4990 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.250664 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.525978 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.561844 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.605809 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.753379 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.770751 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 12:43:20 crc kubenswrapper[4990]: I1203 12:43:20.940120 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.002785 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.126973 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.244800 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.331254 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.359736 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.679698 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.689351 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.784245 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.830042 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 12:43:21 crc kubenswrapper[4990]: I1203 12:43:21.994930 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.145548 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.155424 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.277637 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.339730 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.420712 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.478239 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.543765 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.817200 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 12:43:22 crc kubenswrapper[4990]: I1203 12:43:22.943868 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.168091 4990 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.418684 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.540407 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.630536 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.755678 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.779377 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 12:43:23 crc kubenswrapper[4990]: I1203 12:43:23.855318 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 12:43:24 crc kubenswrapper[4990]: I1203 12:43:24.130189 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 12:43:24 crc kubenswrapper[4990]: I1203 12:43:24.319394 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 12:43:24 crc kubenswrapper[4990]: I1203 12:43:24.371431 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 12:43:24 crc kubenswrapper[4990]: I1203 12:43:24.741099 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 12:43:24 crc kubenswrapper[4990]: I1203 12:43:24.774029 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 12:43:24 crc kubenswrapper[4990]: I1203 12:43:24.895574 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.084195 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.163012 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.185606 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.457091 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.538788 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.671342 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.760669 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.827380 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 12:43:25 crc kubenswrapper[4990]: I1203 12:43:25.874857 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.116472 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.116885 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.117551 4990 scope.go:117] "RemoveContainer" containerID="caadcf267bb9499dd5a413687a3708feda6c22ed87ce1f1f4706f019af7fec70" Dec 03 12:43:26 crc kubenswrapper[4990]: E1203 12:43:26.117758 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-gmsz9_openshift-marketplace(2870df3c-124c-4a2d-8f88-4fec5f65740c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.236470 4990 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.238622 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dbqqz" podStartSLOduration=65.279278261 podStartE2EDuration="3m8.2386044s" podCreationTimestamp="2025-12-03 12:40:18 +0000 UTC" firstStartedPulling="2025-12-03 12:40:23.100829435 +0000 UTC m=+171.242740664" lastFinishedPulling="2025-12-03 12:42:26.060155574 +0000 UTC m=+294.202066803" observedRunningTime="2025-12-03 12:42:36.139564144 +0000 UTC m=+304.281475373" watchObservedRunningTime="2025-12-03 12:43:26.2386044 +0000 UTC m=+354.380515629" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.239145 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sfvnj" podStartSLOduration=81.164951687 podStartE2EDuration="3m6.238747904s" podCreationTimestamp="2025-12-03 12:40:20 +0000 UTC" firstStartedPulling="2025-12-03 12:40:23.035981536 +0000 UTC m=+171.177892765" lastFinishedPulling="2025-12-03 12:42:08.109777753 +0000 UTC m=+276.251688982" observedRunningTime="2025-12-03 12:42:36.1726684 +0000 UTC m=+304.314579639" watchObservedRunningTime="2025-12-03 12:43:26.238747904 +0000 UTC m=+354.380659133" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.239647 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cbxg8" podStartSLOduration=64.287412439 podStartE2EDuration="3m7.239641598s" podCreationTimestamp="2025-12-03 12:40:19 +0000 UTC" firstStartedPulling="2025-12-03 12:40:23.038402329 +0000 UTC m=+171.180313558" lastFinishedPulling="2025-12-03 12:42:25.990631488 +0000 UTC m=+294.132542717" observedRunningTime="2025-12-03 12:42:35.962583049 +0000 UTC m=+304.104494288" watchObservedRunningTime="2025-12-03 12:43:26.239641598 +0000 UTC m=+354.381552827" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.240133 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zn87g" podStartSLOduration=64.258993195 podStartE2EDuration="3m7.240128992s" podCreationTimestamp="2025-12-03 12:40:19 +0000 UTC" firstStartedPulling="2025-12-03 12:40:23.046533508 +0000 UTC m=+171.188444737" lastFinishedPulling="2025-12-03 12:42:26.027669275 +0000 UTC m=+294.169580534" observedRunningTime="2025-12-03 12:42:36.005915075 +0000 UTC m=+304.147826304" watchObservedRunningTime="2025-12-03 12:43:26.240128992 +0000 UTC m=+354.382040221" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.241194 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tprnr" podStartSLOduration=82.056523439 podStartE2EDuration="3m9.24118888s" podCreationTimestamp="2025-12-03 12:40:17 +0000 UTC" firstStartedPulling="2025-12-03 12:40:20.913352169 +0000 UTC m=+169.055263398" lastFinishedPulling="2025-12-03 12:42:08.09801761 +0000 UTC m=+276.239928839" observedRunningTime="2025-12-03 12:42:36.120350231 +0000 UTC m=+304.262261470" watchObservedRunningTime="2025-12-03 12:43:26.24118888 +0000 UTC m=+354.383100109" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.241677 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j98rl" podStartSLOduration=84.276954553 podStartE2EDuration="3m10.241672043s" podCreationTimestamp="2025-12-03 12:40:16 +0000 UTC" firstStartedPulling="2025-12-03 12:40:21.999600129 +0000 UTC m=+170.141511398" lastFinishedPulling="2025-12-03 12:42:07.964317659 +0000 UTC m=+276.106228888" observedRunningTime="2025-12-03 12:42:36.043263571 +0000 UTC m=+304.185174820" watchObservedRunningTime="2025-12-03 12:43:26.241672043 +0000 UTC m=+354.383583272" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.241806 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=90.241803207 podStartE2EDuration="1m30.241803207s" podCreationTimestamp="2025-12-03 12:41:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:42:36.156184084 +0000 UTC m=+304.298095313" watchObservedRunningTime="2025-12-03 12:43:26.241803207 +0000 UTC m=+354.383714436" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.241881 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kjsgh" podStartSLOduration=64.101908449 podStartE2EDuration="3m9.241877789s" podCreationTimestamp="2025-12-03 12:40:17 +0000 UTC" firstStartedPulling="2025-12-03 12:40:20.912653201 +0000 UTC m=+169.054564430" lastFinishedPulling="2025-12-03 12:42:26.052622531 +0000 UTC m=+294.194533770" observedRunningTime="2025-12-03 12:42:36.194374244 +0000 UTC m=+304.336285483" watchObservedRunningTime="2025-12-03 12:43:26.241877789 +0000 UTC m=+354.383789018" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.242305 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jxn9m" podStartSLOduration=83.791730812 podStartE2EDuration="3m10.24230118s" podCreationTimestamp="2025-12-03 12:40:16 +0000 UTC" firstStartedPulling="2025-12-03 12:40:21.991119101 +0000 UTC m=+170.133030330" lastFinishedPulling="2025-12-03 12:42:08.441689459 +0000 UTC m=+276.583600698" observedRunningTime="2025-12-03 12:42:35.985872058 +0000 UTC m=+304.127783307" watchObservedRunningTime="2025-12-03 12:43:26.24230118 +0000 UTC m=+354.384212409" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.242760 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dqrqw","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.242809 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-bd7987fd5-q5rxm"] Dec 03 12:43:26 crc kubenswrapper[4990]: E1203 12:43:26.243014 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" containerName="installer" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243030 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" containerName="installer" Dec 03 12:43:26 crc kubenswrapper[4990]: E1203 12:43:26.243047 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243086 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243221 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" containerName="oauth-openshift" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243223 4990 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243259 4990 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ac72e32a-c8e7-4f8b-a89a-1f1c0ec3a61b" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243236 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1615d35b-f818-40ae-87c2-59fbb9b73fae" containerName="installer" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.243802 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.247830 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248078 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248236 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248271 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248409 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248526 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248553 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248555 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248622 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248668 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248718 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.248874 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.249036 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.252971 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.279837 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.336610 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ccf32a0-8701-4d6f-8c45-d314f98754b3" path="/var/lib/kubelet/pods/2ccf32a0-8701-4d6f-8c45-d314f98754b3/volumes" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.336981 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.339395 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.343687 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.349147 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.352288 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=51.352268302 podStartE2EDuration="51.352268302s" podCreationTimestamp="2025-12-03 12:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:43:26.344209112 +0000 UTC m=+354.486120351" watchObservedRunningTime="2025-12-03 12:43:26.352268302 +0000 UTC m=+354.494179571" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362158 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362216 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-session\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362307 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-error\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362335 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362358 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4zv\" (UniqueName: \"kubernetes.io/projected/7938871f-424b-4f97-934f-886235ba2c2a-kube-api-access-mk4zv\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362385 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-login\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.362418 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363030 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-audit-policies\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363077 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363127 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363170 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363201 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-router-certs\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363231 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7938871f-424b-4f97-934f-886235ba2c2a-audit-dir\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.363254 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-service-ca\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.454445 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464376 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7938871f-424b-4f97-934f-886235ba2c2a-audit-dir\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464473 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-service-ca\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464511 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464536 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7938871f-424b-4f97-934f-886235ba2c2a-audit-dir\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464541 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-session\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464753 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-error\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464870 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464904 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4zv\" (UniqueName: \"kubernetes.io/projected/7938871f-424b-4f97-934f-886235ba2c2a-kube-api-access-mk4zv\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464930 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-login\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.464999 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.465054 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-audit-policies\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.465324 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-service-ca\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.465699 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.465774 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.465840 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.465868 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-router-certs\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.467976 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.468517 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-audit-policies\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.473396 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.476988 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.478628 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-error\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.480824 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-session\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.481534 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.482062 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-login\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.487775 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-router-certs\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.487775 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4zv\" (UniqueName: \"kubernetes.io/projected/7938871f-424b-4f97-934f-886235ba2c2a-kube-api-access-mk4zv\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.494383 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.517869 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7938871f-424b-4f97-934f-886235ba2c2a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-bd7987fd5-q5rxm\" (UID: \"7938871f-424b-4f97-934f-886235ba2c2a\") " pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.570495 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.590484 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.751609 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 12:43:26 crc kubenswrapper[4990]: I1203 12:43:26.809415 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 12:43:27 crc kubenswrapper[4990]: I1203 12:43:26.942641 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 12:43:27 crc kubenswrapper[4990]: I1203 12:43:27.285678 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 12:43:27 crc kubenswrapper[4990]: I1203 12:43:27.600018 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 12:43:27 crc kubenswrapper[4990]: I1203 12:43:27.869968 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 12:43:27 crc kubenswrapper[4990]: I1203 12:43:27.942228 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 12:43:28 crc kubenswrapper[4990]: I1203 12:43:28.266170 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 12:43:28 crc kubenswrapper[4990]: I1203 12:43:28.366364 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 12:43:28 crc kubenswrapper[4990]: I1203 12:43:28.583786 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 12:43:28 crc kubenswrapper[4990]: I1203 12:43:28.787289 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 12:43:28 crc kubenswrapper[4990]: I1203 12:43:28.984011 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 12:43:29 crc kubenswrapper[4990]: I1203 12:43:29.125637 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 12:43:29 crc kubenswrapper[4990]: I1203 12:43:29.819354 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 12:43:29 crc kubenswrapper[4990]: I1203 12:43:29.882512 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 12:43:29 crc kubenswrapper[4990]: I1203 12:43:29.894207 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 12:43:30 crc kubenswrapper[4990]: I1203 12:43:30.380996 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 12:43:30 crc kubenswrapper[4990]: E1203 12:43:30.389015 4990 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 12:43:30 crc kubenswrapper[4990]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-bd7987fd5-q5rxm_openshift-authentication_7938871f-424b-4f97-934f-886235ba2c2a_0(6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb): error adding pod openshift-authentication_oauth-openshift-bd7987fd5-q5rxm to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb" Netns:"/var/run/netns/c7e9fdd5-51d7-49c5-9898-7e397be16b6d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-bd7987fd5-q5rxm;K8S_POD_INFRA_CONTAINER_ID=6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb;K8S_POD_UID=7938871f-424b-4f97-934f-886235ba2c2a" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm] networking: Multus: [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm/7938871f-424b-4f97-934f-886235ba2c2a]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-bd7987fd5-q5rxm in out of cluster comm: pod "oauth-openshift-bd7987fd5-q5rxm" not found Dec 03 12:43:30 crc kubenswrapper[4990]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 12:43:30 crc kubenswrapper[4990]: > Dec 03 12:43:30 crc kubenswrapper[4990]: E1203 12:43:30.389113 4990 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 12:43:30 crc kubenswrapper[4990]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-bd7987fd5-q5rxm_openshift-authentication_7938871f-424b-4f97-934f-886235ba2c2a_0(6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb): error adding pod openshift-authentication_oauth-openshift-bd7987fd5-q5rxm to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb" Netns:"/var/run/netns/c7e9fdd5-51d7-49c5-9898-7e397be16b6d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-bd7987fd5-q5rxm;K8S_POD_INFRA_CONTAINER_ID=6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb;K8S_POD_UID=7938871f-424b-4f97-934f-886235ba2c2a" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm] networking: Multus: [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm/7938871f-424b-4f97-934f-886235ba2c2a]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-bd7987fd5-q5rxm in out of cluster comm: pod "oauth-openshift-bd7987fd5-q5rxm" not found Dec 03 12:43:30 crc kubenswrapper[4990]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 12:43:30 crc kubenswrapper[4990]: > pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:30 crc kubenswrapper[4990]: E1203 12:43:30.389146 4990 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 03 12:43:30 crc kubenswrapper[4990]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-bd7987fd5-q5rxm_openshift-authentication_7938871f-424b-4f97-934f-886235ba2c2a_0(6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb): error adding pod openshift-authentication_oauth-openshift-bd7987fd5-q5rxm to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb" Netns:"/var/run/netns/c7e9fdd5-51d7-49c5-9898-7e397be16b6d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-bd7987fd5-q5rxm;K8S_POD_INFRA_CONTAINER_ID=6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb;K8S_POD_UID=7938871f-424b-4f97-934f-886235ba2c2a" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm] networking: Multus: [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm/7938871f-424b-4f97-934f-886235ba2c2a]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-bd7987fd5-q5rxm in out of cluster comm: pod "oauth-openshift-bd7987fd5-q5rxm" not found Dec 03 12:43:30 crc kubenswrapper[4990]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 12:43:30 crc kubenswrapper[4990]: > pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:30 crc kubenswrapper[4990]: E1203 12:43:30.389208 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-bd7987fd5-q5rxm_openshift-authentication(7938871f-424b-4f97-934f-886235ba2c2a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-bd7987fd5-q5rxm_openshift-authentication(7938871f-424b-4f97-934f-886235ba2c2a)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-bd7987fd5-q5rxm_openshift-authentication_7938871f-424b-4f97-934f-886235ba2c2a_0(6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb): error adding pod openshift-authentication_oauth-openshift-bd7987fd5-q5rxm to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb\\\" Netns:\\\"/var/run/netns/c7e9fdd5-51d7-49c5-9898-7e397be16b6d\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-bd7987fd5-q5rxm;K8S_POD_INFRA_CONTAINER_ID=6b49963cfcd93d59bc6ccd4efe5399a5fb80696600ca4632e6843d1a2424d8eb;K8S_POD_UID=7938871f-424b-4f97-934f-886235ba2c2a\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm] networking: Multus: [openshift-authentication/oauth-openshift-bd7987fd5-q5rxm/7938871f-424b-4f97-934f-886235ba2c2a]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-bd7987fd5-q5rxm in out of cluster comm: pod \\\"oauth-openshift-bd7987fd5-q5rxm\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" podUID="7938871f-424b-4f97-934f-886235ba2c2a" Dec 03 12:43:31 crc kubenswrapper[4990]: I1203 12:43:31.096599 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 12:43:31 crc kubenswrapper[4990]: I1203 12:43:31.301949 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 12:43:31 crc kubenswrapper[4990]: I1203 12:43:31.494744 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 12:43:32 crc kubenswrapper[4990]: I1203 12:43:32.103299 4990 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 12:43:32 crc kubenswrapper[4990]: I1203 12:43:32.103574 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://aff99b436d50dc5c6e07784235b5da37c236aea6f8afacc7939bb83e42234f9e" gracePeriod=5 Dec 03 12:43:33 crc kubenswrapper[4990]: I1203 12:43:33.046888 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 12:43:33 crc kubenswrapper[4990]: I1203 12:43:33.286787 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:43:33 crc kubenswrapper[4990]: I1203 12:43:33.286849 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:43:37 crc kubenswrapper[4990]: I1203 12:43:37.775920 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 12:43:37 crc kubenswrapper[4990]: I1203 12:43:37.776423 4990 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="aff99b436d50dc5c6e07784235b5da37c236aea6f8afacc7939bb83e42234f9e" exitCode=137 Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.275248 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.275397 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.395967 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396093 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396121 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396163 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396221 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396329 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396640 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396681 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396700 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396877 4990 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396902 4990 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396918 4990 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.396939 4990 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.404076 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.498559 4990 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.783240 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.783306 4990 scope.go:117] "RemoveContainer" containerID="aff99b436d50dc5c6e07784235b5da37c236aea6f8afacc7939bb83e42234f9e" Dec 03 12:43:38 crc kubenswrapper[4990]: I1203 12:43:38.783369 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 12:43:39 crc kubenswrapper[4990]: I1203 12:43:39.264799 4990 scope.go:117] "RemoveContainer" containerID="caadcf267bb9499dd5a413687a3708feda6c22ed87ce1f1f4706f019af7fec70" Dec 03 12:43:39 crc kubenswrapper[4990]: I1203 12:43:39.791816 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gmsz9_2870df3c-124c-4a2d-8f88-4fec5f65740c/marketplace-operator/1.log" Dec 03 12:43:39 crc kubenswrapper[4990]: I1203 12:43:39.792155 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerStarted","Data":"cdbd7460ad4601847233699deb0e5f3d3d63e279d3c37733744f03a1406cf537"} Dec 03 12:43:39 crc kubenswrapper[4990]: I1203 12:43:39.793423 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:43:39 crc kubenswrapper[4990]: I1203 12:43:39.794434 4990 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gmsz9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 03 12:43:39 crc kubenswrapper[4990]: I1203 12:43:39.796582 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.240429 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.271146 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.271470 4990 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.280817 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.280861 4990 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="291a3585-d1a3-494e-9e44-9d60f4341d36" Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.285689 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.285732 4990 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="291a3585-d1a3-494e-9e44-9d60f4341d36" Dec 03 12:43:40 crc kubenswrapper[4990]: I1203 12:43:40.809980 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:43:41 crc kubenswrapper[4990]: I1203 12:43:41.636557 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 12:43:42 crc kubenswrapper[4990]: I1203 12:43:42.263868 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:42 crc kubenswrapper[4990]: I1203 12:43:42.267254 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:42 crc kubenswrapper[4990]: I1203 12:43:42.491201 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-bd7987fd5-q5rxm"] Dec 03 12:43:42 crc kubenswrapper[4990]: W1203 12:43:42.509196 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7938871f_424b_4f97_934f_886235ba2c2a.slice/crio-a1fcdc719b3646de5d27d8423397fb443495ca2bd38f7dea133628edcb95616a WatchSource:0}: Error finding container a1fcdc719b3646de5d27d8423397fb443495ca2bd38f7dea133628edcb95616a: Status 404 returned error can't find the container with id a1fcdc719b3646de5d27d8423397fb443495ca2bd38f7dea133628edcb95616a Dec 03 12:43:42 crc kubenswrapper[4990]: I1203 12:43:42.822860 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" event={"ID":"7938871f-424b-4f97-934f-886235ba2c2a","Type":"ContainerStarted","Data":"6ba6cf4dabf3f0f398efafb6df1b035b83dae68ff13cb8a105698bc13e65226a"} Dec 03 12:43:42 crc kubenswrapper[4990]: I1203 12:43:42.823217 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:42 crc kubenswrapper[4990]: I1203 12:43:42.823233 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" event={"ID":"7938871f-424b-4f97-934f-886235ba2c2a","Type":"ContainerStarted","Data":"a1fcdc719b3646de5d27d8423397fb443495ca2bd38f7dea133628edcb95616a"} Dec 03 12:43:43 crc kubenswrapper[4990]: I1203 12:43:43.171757 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" Dec 03 12:43:43 crc kubenswrapper[4990]: I1203 12:43:43.199660 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-bd7987fd5-q5rxm" podStartSLOduration=138.199633264 podStartE2EDuration="2m18.199633264s" podCreationTimestamp="2025-12-03 12:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:43:42.852585584 +0000 UTC m=+370.994496833" watchObservedRunningTime="2025-12-03 12:43:43.199633264 +0000 UTC m=+371.341544493" Dec 03 12:43:54 crc kubenswrapper[4990]: I1203 12:43:54.476097 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 12:43:55 crc kubenswrapper[4990]: I1203 12:43:55.678488 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 12:43:56 crc kubenswrapper[4990]: I1203 12:43:56.143942 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 12:43:56 crc kubenswrapper[4990]: I1203 12:43:56.376314 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.365617 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.404628 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqdjl"] Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.405386 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" podUID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" containerName="controller-manager" containerID="cri-o://b60c1fc2097278682f43c5ca3d540f8fc5a3c85a967e384607642fa270c602a7" gracePeriod=30 Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.512901 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v"] Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.513835 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" podUID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" containerName="route-controller-manager" containerID="cri-o://a503f87f46275060687617261477a76d0477648eeedf9015ff7f0d07d2f48977" gracePeriod=30 Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.936732 4990 generic.go:334] "Generic (PLEG): container finished" podID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" containerID="a503f87f46275060687617261477a76d0477648eeedf9015ff7f0d07d2f48977" exitCode=0 Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.936837 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" event={"ID":"1a2c5a8d-266b-441a-9977-16b138ebcc6f","Type":"ContainerDied","Data":"a503f87f46275060687617261477a76d0477648eeedf9015ff7f0d07d2f48977"} Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.938887 4990 generic.go:334] "Generic (PLEG): container finished" podID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" containerID="b60c1fc2097278682f43c5ca3d540f8fc5a3c85a967e384607642fa270c602a7" exitCode=0 Dec 03 12:43:59 crc kubenswrapper[4990]: I1203 12:43:59.938969 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" event={"ID":"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906","Type":"ContainerDied","Data":"b60c1fc2097278682f43c5ca3d540f8fc5a3c85a967e384607642fa270c602a7"} Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.396493 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.548583 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwtdt\" (UniqueName: \"kubernetes.io/projected/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-kube-api-access-jwtdt\") pod \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.549248 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-serving-cert\") pod \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.549312 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-config\") pod \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.549364 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-proxy-ca-bundles\") pod \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.549428 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-client-ca\") pod \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\" (UID: \"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.550490 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-client-ca" (OuterVolumeSpecName: "client-ca") pod "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" (UID: "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.550547 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" (UID: "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.550665 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-config" (OuterVolumeSpecName: "config") pod "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" (UID: "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.600067 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-kube-api-access-jwtdt" (OuterVolumeSpecName: "kube-api-access-jwtdt") pod "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" (UID: "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906"). InnerVolumeSpecName "kube-api-access-jwtdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.600804 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" (UID: "7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.601955 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.650826 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.650893 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwtdt\" (UniqueName: \"kubernetes.io/projected/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-kube-api-access-jwtdt\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.650911 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.650923 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.650934 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.752397 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-config\") pod \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.752501 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-client-ca\") pod \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.752637 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pl7s\" (UniqueName: \"kubernetes.io/projected/1a2c5a8d-266b-441a-9977-16b138ebcc6f-kube-api-access-7pl7s\") pod \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.752710 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a2c5a8d-266b-441a-9977-16b138ebcc6f-serving-cert\") pod \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\" (UID: \"1a2c5a8d-266b-441a-9977-16b138ebcc6f\") " Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.754883 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-client-ca" (OuterVolumeSpecName: "client-ca") pod "1a2c5a8d-266b-441a-9977-16b138ebcc6f" (UID: "1a2c5a8d-266b-441a-9977-16b138ebcc6f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.755129 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-config" (OuterVolumeSpecName: "config") pod "1a2c5a8d-266b-441a-9977-16b138ebcc6f" (UID: "1a2c5a8d-266b-441a-9977-16b138ebcc6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.757913 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2c5a8d-266b-441a-9977-16b138ebcc6f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1a2c5a8d-266b-441a-9977-16b138ebcc6f" (UID: "1a2c5a8d-266b-441a-9977-16b138ebcc6f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.759544 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2c5a8d-266b-441a-9977-16b138ebcc6f-kube-api-access-7pl7s" (OuterVolumeSpecName: "kube-api-access-7pl7s") pod "1a2c5a8d-266b-441a-9977-16b138ebcc6f" (UID: "1a2c5a8d-266b-441a-9977-16b138ebcc6f"). InnerVolumeSpecName "kube-api-access-7pl7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.854832 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a2c5a8d-266b-441a-9977-16b138ebcc6f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.854887 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.854900 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a2c5a8d-266b-441a-9977-16b138ebcc6f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.854912 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pl7s\" (UniqueName: \"kubernetes.io/projected/1a2c5a8d-266b-441a-9977-16b138ebcc6f-kube-api-access-7pl7s\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.948207 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" event={"ID":"7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906","Type":"ContainerDied","Data":"ef29f59bdcbbcb5a0358254e14da115cd36feb4d3c25acdaedf3d642824a78d2"} Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.948279 4990 scope.go:117] "RemoveContainer" containerID="b60c1fc2097278682f43c5ca3d540f8fc5a3c85a967e384607642fa270c602a7" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.948275 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqdjl" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.957059 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" event={"ID":"1a2c5a8d-266b-441a-9977-16b138ebcc6f","Type":"ContainerDied","Data":"b8969ac0ce41a706aab7a8d93b5b272a289b1a9c351d1dd138a415719c0d8d58"} Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.957124 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v" Dec 03 12:44:00 crc kubenswrapper[4990]: I1203 12:44:00.979286 4990 scope.go:117] "RemoveContainer" containerID="a503f87f46275060687617261477a76d0477648eeedf9015ff7f0d07d2f48977" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.001802 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.020456 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-25q7v"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.026983 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqdjl"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.031901 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqdjl"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246138 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4"] Dec 03 12:44:01 crc kubenswrapper[4990]: E1203 12:44:01.246618 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" containerName="route-controller-manager" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246641 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" containerName="route-controller-manager" Dec 03 12:44:01 crc kubenswrapper[4990]: E1203 12:44:01.246656 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" containerName="controller-manager" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246666 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" containerName="controller-manager" Dec 03 12:44:01 crc kubenswrapper[4990]: E1203 12:44:01.246683 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246692 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246833 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246859 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" containerName="route-controller-manager" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.246872 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" containerName="controller-manager" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.247506 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.250121 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.250479 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.251317 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.253124 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.253127 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.253208 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.254417 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.258123 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.258218 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.258431 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.258554 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.258590 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.258787 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.261506 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.269778 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.276792 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc"] Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.280674 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.363424 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwqq2\" (UniqueName: \"kubernetes.io/projected/a97f6cfa-be32-475c-9b42-d9511821cde0-kube-api-access-zwqq2\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.363842 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-client-ca\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.363879 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj9qj\" (UniqueName: \"kubernetes.io/projected/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-kube-api-access-sj9qj\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.363916 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-serving-cert\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.364132 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a97f6cfa-be32-475c-9b42-d9511821cde0-serving-cert\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.364239 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-client-ca\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.364449 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-config\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.364605 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-proxy-ca-bundles\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.364667 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-config\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466530 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-config\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466589 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-proxy-ca-bundles\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466613 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-config\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466642 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwqq2\" (UniqueName: \"kubernetes.io/projected/a97f6cfa-be32-475c-9b42-d9511821cde0-kube-api-access-zwqq2\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466681 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-client-ca\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466713 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj9qj\" (UniqueName: \"kubernetes.io/projected/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-kube-api-access-sj9qj\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466759 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-serving-cert\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466795 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a97f6cfa-be32-475c-9b42-d9511821cde0-serving-cert\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.466840 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-client-ca\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.468639 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-client-ca\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.468903 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-config\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.469303 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-proxy-ca-bundles\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.469408 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-config\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.470342 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-client-ca\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.472980 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a97f6cfa-be32-475c-9b42-d9511821cde0-serving-cert\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.475270 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-serving-cert\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.491981 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwqq2\" (UniqueName: \"kubernetes.io/projected/a97f6cfa-be32-475c-9b42-d9511821cde0-kube-api-access-zwqq2\") pod \"controller-manager-5ff74b54fd-4xpnc\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.494665 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj9qj\" (UniqueName: \"kubernetes.io/projected/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-kube-api-access-sj9qj\") pod \"route-controller-manager-7d94b58c74-qp9k4\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.571949 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.589749 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.808966 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc"] Dec 03 12:44:01 crc kubenswrapper[4990]: W1203 12:44:01.828354 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda97f6cfa_be32_475c_9b42_d9511821cde0.slice/crio-d732c072fd5b36f480086e9bb5b60b97d582abd920ada850329ea05ffbfa461a WatchSource:0}: Error finding container d732c072fd5b36f480086e9bb5b60b97d582abd920ada850329ea05ffbfa461a: Status 404 returned error can't find the container with id d732c072fd5b36f480086e9bb5b60b97d582abd920ada850329ea05ffbfa461a Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.841067 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4"] Dec 03 12:44:01 crc kubenswrapper[4990]: W1203 12:44:01.861928 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5530dfa_f256_4f95_b6bd_cfc96cdd341e.slice/crio-7d26e1ea9d4a33af2db95789a4d7a4d89a2506f6dab34bf43d9f4a7691144773 WatchSource:0}: Error finding container 7d26e1ea9d4a33af2db95789a4d7a4d89a2506f6dab34bf43d9f4a7691144773: Status 404 returned error can't find the container with id 7d26e1ea9d4a33af2db95789a4d7a4d89a2506f6dab34bf43d9f4a7691144773 Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.967183 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" event={"ID":"c5530dfa-f256-4f95-b6bd-cfc96cdd341e","Type":"ContainerStarted","Data":"7d26e1ea9d4a33af2db95789a4d7a4d89a2506f6dab34bf43d9f4a7691144773"} Dec 03 12:44:01 crc kubenswrapper[4990]: I1203 12:44:01.968310 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" event={"ID":"a97f6cfa-be32-475c-9b42-d9511821cde0","Type":"ContainerStarted","Data":"d732c072fd5b36f480086e9bb5b60b97d582abd920ada850329ea05ffbfa461a"} Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.275987 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a2c5a8d-266b-441a-9977-16b138ebcc6f" path="/var/lib/kubelet/pods/1a2c5a8d-266b-441a-9977-16b138ebcc6f/volumes" Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.277104 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906" path="/var/lib/kubelet/pods/7c24e0b4-310e-40c5-8ba6-c2fcd5a6b906/volumes" Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.828781 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.977680 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" event={"ID":"a97f6cfa-be32-475c-9b42-d9511821cde0","Type":"ContainerStarted","Data":"ec28d3daf702983154dae08c7d0e3f313ec6aba7c392726f7b0893345e64ae3c"} Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.978254 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.980346 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" event={"ID":"c5530dfa-f256-4f95-b6bd-cfc96cdd341e","Type":"ContainerStarted","Data":"49776880c2d2a4730e956acf6d81a755e470c8d4c6ac3b50d8adf59d74577e7b"} Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.980638 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.989714 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:02 crc kubenswrapper[4990]: I1203 12:44:02.989845 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:03 crc kubenswrapper[4990]: I1203 12:44:03.012660 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" podStartSLOduration=4.012631072 podStartE2EDuration="4.012631072s" podCreationTimestamp="2025-12-03 12:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:03.005985181 +0000 UTC m=+391.147896420" watchObservedRunningTime="2025-12-03 12:44:03.012631072 +0000 UTC m=+391.154542301" Dec 03 12:44:03 crc kubenswrapper[4990]: I1203 12:44:03.060596 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" podStartSLOduration=4.060579676 podStartE2EDuration="4.060579676s" podCreationTimestamp="2025-12-03 12:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:03.059143467 +0000 UTC m=+391.201054696" watchObservedRunningTime="2025-12-03 12:44:03.060579676 +0000 UTC m=+391.202490905" Dec 03 12:44:03 crc kubenswrapper[4990]: I1203 12:44:03.286651 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:44:03 crc kubenswrapper[4990]: I1203 12:44:03.286753 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:44:09 crc kubenswrapper[4990]: I1203 12:44:09.387505 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 12:44:15 crc kubenswrapper[4990]: I1203 12:44:15.420041 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc"] Dec 03 12:44:15 crc kubenswrapper[4990]: I1203 12:44:15.420832 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" podUID="a97f6cfa-be32-475c-9b42-d9511821cde0" containerName="controller-manager" containerID="cri-o://ec28d3daf702983154dae08c7d0e3f313ec6aba7c392726f7b0893345e64ae3c" gracePeriod=30 Dec 03 12:44:15 crc kubenswrapper[4990]: I1203 12:44:15.445432 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4"] Dec 03 12:44:15 crc kubenswrapper[4990]: I1203 12:44:15.445689 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" podUID="c5530dfa-f256-4f95-b6bd-cfc96cdd341e" containerName="route-controller-manager" containerID="cri-o://49776880c2d2a4730e956acf6d81a755e470c8d4c6ac3b50d8adf59d74577e7b" gracePeriod=30 Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.080495 4990 generic.go:334] "Generic (PLEG): container finished" podID="c5530dfa-f256-4f95-b6bd-cfc96cdd341e" containerID="49776880c2d2a4730e956acf6d81a755e470c8d4c6ac3b50d8adf59d74577e7b" exitCode=0 Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.080586 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" event={"ID":"c5530dfa-f256-4f95-b6bd-cfc96cdd341e","Type":"ContainerDied","Data":"49776880c2d2a4730e956acf6d81a755e470c8d4c6ac3b50d8adf59d74577e7b"} Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.083673 4990 generic.go:334] "Generic (PLEG): container finished" podID="a97f6cfa-be32-475c-9b42-d9511821cde0" containerID="ec28d3daf702983154dae08c7d0e3f313ec6aba7c392726f7b0893345e64ae3c" exitCode=0 Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.083740 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" event={"ID":"a97f6cfa-be32-475c-9b42-d9511821cde0","Type":"ContainerDied","Data":"ec28d3daf702983154dae08c7d0e3f313ec6aba7c392726f7b0893345e64ae3c"} Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.688885 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.719847 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq"] Dec 03 12:44:16 crc kubenswrapper[4990]: E1203 12:44:16.720156 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5530dfa-f256-4f95-b6bd-cfc96cdd341e" containerName="route-controller-manager" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.720176 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5530dfa-f256-4f95-b6bd-cfc96cdd341e" containerName="route-controller-manager" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.720288 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5530dfa-f256-4f95-b6bd-cfc96cdd341e" containerName="route-controller-manager" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.720815 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.729009 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq"] Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.760368 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.857113 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-serving-cert\") pod \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.857445 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-config\") pod \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.857502 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-client-ca\") pod \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.857533 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj9qj\" (UniqueName: \"kubernetes.io/projected/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-kube-api-access-sj9qj\") pod \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\" (UID: \"c5530dfa-f256-4f95-b6bd-cfc96cdd341e\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.858114 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-client-ca\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.858154 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-config\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.858203 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrwck\" (UniqueName: \"kubernetes.io/projected/9c20895f-e512-47fe-b62b-d7d47080bbd1-kube-api-access-qrwck\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.858227 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c20895f-e512-47fe-b62b-d7d47080bbd1-serving-cert\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.859000 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-client-ca" (OuterVolumeSpecName: "client-ca") pod "c5530dfa-f256-4f95-b6bd-cfc96cdd341e" (UID: "c5530dfa-f256-4f95-b6bd-cfc96cdd341e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.860059 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-config" (OuterVolumeSpecName: "config") pod "c5530dfa-f256-4f95-b6bd-cfc96cdd341e" (UID: "c5530dfa-f256-4f95-b6bd-cfc96cdd341e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.865797 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-kube-api-access-sj9qj" (OuterVolumeSpecName: "kube-api-access-sj9qj") pod "c5530dfa-f256-4f95-b6bd-cfc96cdd341e" (UID: "c5530dfa-f256-4f95-b6bd-cfc96cdd341e"). InnerVolumeSpecName "kube-api-access-sj9qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.866252 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c5530dfa-f256-4f95-b6bd-cfc96cdd341e" (UID: "c5530dfa-f256-4f95-b6bd-cfc96cdd341e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.959560 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwqq2\" (UniqueName: \"kubernetes.io/projected/a97f6cfa-be32-475c-9b42-d9511821cde0-kube-api-access-zwqq2\") pod \"a97f6cfa-be32-475c-9b42-d9511821cde0\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.959667 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-client-ca\") pod \"a97f6cfa-be32-475c-9b42-d9511821cde0\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.959759 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-proxy-ca-bundles\") pod \"a97f6cfa-be32-475c-9b42-d9511821cde0\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.959827 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a97f6cfa-be32-475c-9b42-d9511821cde0-serving-cert\") pod \"a97f6cfa-be32-475c-9b42-d9511821cde0\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.959920 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-config\") pod \"a97f6cfa-be32-475c-9b42-d9511821cde0\" (UID: \"a97f6cfa-be32-475c-9b42-d9511821cde0\") " Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960163 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-config\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960302 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrwck\" (UniqueName: \"kubernetes.io/projected/9c20895f-e512-47fe-b62b-d7d47080bbd1-kube-api-access-qrwck\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960345 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c20895f-e512-47fe-b62b-d7d47080bbd1-serving-cert\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960433 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-client-ca\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960534 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960568 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960588 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.960607 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj9qj\" (UniqueName: \"kubernetes.io/projected/c5530dfa-f256-4f95-b6bd-cfc96cdd341e-kube-api-access-sj9qj\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.961421 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-client-ca" (OuterVolumeSpecName: "client-ca") pod "a97f6cfa-be32-475c-9b42-d9511821cde0" (UID: "a97f6cfa-be32-475c-9b42-d9511821cde0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.962384 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a97f6cfa-be32-475c-9b42-d9511821cde0" (UID: "a97f6cfa-be32-475c-9b42-d9511821cde0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.962559 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-config" (OuterVolumeSpecName: "config") pod "a97f6cfa-be32-475c-9b42-d9511821cde0" (UID: "a97f6cfa-be32-475c-9b42-d9511821cde0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.964010 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-client-ca\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.964505 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-config\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.966314 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a97f6cfa-be32-475c-9b42-d9511821cde0-kube-api-access-zwqq2" (OuterVolumeSpecName: "kube-api-access-zwqq2") pod "a97f6cfa-be32-475c-9b42-d9511821cde0" (UID: "a97f6cfa-be32-475c-9b42-d9511821cde0"). InnerVolumeSpecName "kube-api-access-zwqq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.966425 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a97f6cfa-be32-475c-9b42-d9511821cde0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a97f6cfa-be32-475c-9b42-d9511821cde0" (UID: "a97f6cfa-be32-475c-9b42-d9511821cde0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.973879 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c20895f-e512-47fe-b62b-d7d47080bbd1-serving-cert\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:16 crc kubenswrapper[4990]: I1203 12:44:16.982595 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrwck\" (UniqueName: \"kubernetes.io/projected/9c20895f-e512-47fe-b62b-d7d47080bbd1-kube-api-access-qrwck\") pod \"route-controller-manager-646bddd5f-drhqq\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.061511 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.061565 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwqq2\" (UniqueName: \"kubernetes.io/projected/a97f6cfa-be32-475c-9b42-d9511821cde0-kube-api-access-zwqq2\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.061581 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.061595 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a97f6cfa-be32-475c-9b42-d9511821cde0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.061606 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a97f6cfa-be32-475c-9b42-d9511821cde0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.071168 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.094026 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" event={"ID":"a97f6cfa-be32-475c-9b42-d9511821cde0","Type":"ContainerDied","Data":"d732c072fd5b36f480086e9bb5b60b97d582abd920ada850329ea05ffbfa461a"} Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.094202 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.094263 4990 scope.go:117] "RemoveContainer" containerID="ec28d3daf702983154dae08c7d0e3f313ec6aba7c392726f7b0893345e64ae3c" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.097043 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" event={"ID":"c5530dfa-f256-4f95-b6bd-cfc96cdd341e","Type":"ContainerDied","Data":"7d26e1ea9d4a33af2db95789a4d7a4d89a2506f6dab34bf43d9f4a7691144773"} Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.097352 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.133169 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc"] Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.135163 4990 scope.go:117] "RemoveContainer" containerID="49776880c2d2a4730e956acf6d81a755e470c8d4c6ac3b50d8adf59d74577e7b" Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.142196 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5ff74b54fd-4xpnc"] Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.149961 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4"] Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.153680 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d94b58c74-qp9k4"] Dec 03 12:44:17 crc kubenswrapper[4990]: I1203 12:44:17.402482 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq"] Dec 03 12:44:17 crc kubenswrapper[4990]: W1203 12:44:17.410339 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c20895f_e512_47fe_b62b_d7d47080bbd1.slice/crio-716b7c4617c87694fb013050f2d5ea470a35d44584d547c22cc90e8fcda1ac08 WatchSource:0}: Error finding container 716b7c4617c87694fb013050f2d5ea470a35d44584d547c22cc90e8fcda1ac08: Status 404 returned error can't find the container with id 716b7c4617c87694fb013050f2d5ea470a35d44584d547c22cc90e8fcda1ac08 Dec 03 12:44:18 crc kubenswrapper[4990]: I1203 12:44:18.106249 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" event={"ID":"9c20895f-e512-47fe-b62b-d7d47080bbd1","Type":"ContainerStarted","Data":"716b7c4617c87694fb013050f2d5ea470a35d44584d547c22cc90e8fcda1ac08"} Dec 03 12:44:18 crc kubenswrapper[4990]: I1203 12:44:18.277004 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a97f6cfa-be32-475c-9b42-d9511821cde0" path="/var/lib/kubelet/pods/a97f6cfa-be32-475c-9b42-d9511821cde0/volumes" Dec 03 12:44:18 crc kubenswrapper[4990]: I1203 12:44:18.279002 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5530dfa-f256-4f95-b6bd-cfc96cdd341e" path="/var/lib/kubelet/pods/c5530dfa-f256-4f95-b6bd-cfc96cdd341e/volumes" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.260376 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-m4xsj"] Dec 03 12:44:19 crc kubenswrapper[4990]: E1203 12:44:19.260652 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97f6cfa-be32-475c-9b42-d9511821cde0" containerName="controller-manager" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.260665 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97f6cfa-be32-475c-9b42-d9511821cde0" containerName="controller-manager" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.260783 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="a97f6cfa-be32-475c-9b42-d9511821cde0" containerName="controller-manager" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.261216 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.264031 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.264499 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.264728 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.264935 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.264958 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.265076 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.271999 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.278750 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-m4xsj"] Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.292910 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-config\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.293147 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f57f24a-6040-41ff-9077-75038fc76de4-serving-cert\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.293309 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wdp6\" (UniqueName: \"kubernetes.io/projected/9f57f24a-6040-41ff-9077-75038fc76de4-kube-api-access-6wdp6\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.293470 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-client-ca\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.293594 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-proxy-ca-bundles\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.395440 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-config\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.395564 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f57f24a-6040-41ff-9077-75038fc76de4-serving-cert\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.395598 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wdp6\" (UniqueName: \"kubernetes.io/projected/9f57f24a-6040-41ff-9077-75038fc76de4-kube-api-access-6wdp6\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.395633 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-client-ca\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.395660 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-proxy-ca-bundles\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.396975 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-client-ca\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.397131 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-config\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.397187 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-proxy-ca-bundles\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.406842 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f57f24a-6040-41ff-9077-75038fc76de4-serving-cert\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.433093 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wdp6\" (UniqueName: \"kubernetes.io/projected/9f57f24a-6040-41ff-9077-75038fc76de4-kube-api-access-6wdp6\") pod \"controller-manager-55b446ff67-m4xsj\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:19 crc kubenswrapper[4990]: I1203 12:44:19.581287 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:20 crc kubenswrapper[4990]: I1203 12:44:20.033013 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-m4xsj"] Dec 03 12:44:20 crc kubenswrapper[4990]: I1203 12:44:20.121037 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" event={"ID":"9f57f24a-6040-41ff-9077-75038fc76de4","Type":"ContainerStarted","Data":"c2087b83781f8ebaae246dba664a038e78f3339fc2233a82eb02b8395d6b976a"} Dec 03 12:44:20 crc kubenswrapper[4990]: I1203 12:44:20.123251 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" event={"ID":"9c20895f-e512-47fe-b62b-d7d47080bbd1","Type":"ContainerStarted","Data":"7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70"} Dec 03 12:44:20 crc kubenswrapper[4990]: I1203 12:44:20.125418 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:20 crc kubenswrapper[4990]: I1203 12:44:20.150935 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" podStartSLOduration=5.150914654 podStartE2EDuration="5.150914654s" podCreationTimestamp="2025-12-03 12:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:20.15035479 +0000 UTC m=+408.292266029" watchObservedRunningTime="2025-12-03 12:44:20.150914654 +0000 UTC m=+408.292825903" Dec 03 12:44:20 crc kubenswrapper[4990]: I1203 12:44:20.399288 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:21 crc kubenswrapper[4990]: I1203 12:44:21.132034 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" event={"ID":"9f57f24a-6040-41ff-9077-75038fc76de4","Type":"ContainerStarted","Data":"330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2"} Dec 03 12:44:21 crc kubenswrapper[4990]: I1203 12:44:21.151034 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" podStartSLOduration=6.1510078870000005 podStartE2EDuration="6.151007887s" podCreationTimestamp="2025-12-03 12:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:21.149549389 +0000 UTC m=+409.291460618" watchObservedRunningTime="2025-12-03 12:44:21.151007887 +0000 UTC m=+409.292919116" Dec 03 12:44:22 crc kubenswrapper[4990]: I1203 12:44:22.138552 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:22 crc kubenswrapper[4990]: I1203 12:44:22.143491 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.407880 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9hdr5"] Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.409457 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.422916 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9hdr5"] Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.545550 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-bound-sa-token\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.545873 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzhkc\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-kube-api-access-fzhkc\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.545975 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-registry-certificates\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.546105 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-registry-tls\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.546959 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.547079 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.547173 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-trusted-ca\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.547304 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.600093 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649461 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-registry-tls\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649678 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649712 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649739 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-trusted-ca\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649807 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-bound-sa-token\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649840 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzhkc\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-kube-api-access-fzhkc\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.649866 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-registry-certificates\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.651310 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.651852 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-trusted-ca\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.652207 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-registry-certificates\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.658078 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-registry-tls\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.658481 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.674502 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-bound-sa-token\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.679178 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzhkc\" (UniqueName: \"kubernetes.io/projected/f4b5755f-a0e8-4104-b4d1-5b35423eaa83-kube-api-access-fzhkc\") pod \"image-registry-66df7c8f76-9hdr5\" (UID: \"f4b5755f-a0e8-4104-b4d1-5b35423eaa83\") " pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:27 crc kubenswrapper[4990]: I1203 12:44:27.730028 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:28 crc kubenswrapper[4990]: I1203 12:44:28.222548 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9hdr5"] Dec 03 12:44:29 crc kubenswrapper[4990]: I1203 12:44:29.191876 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" event={"ID":"f4b5755f-a0e8-4104-b4d1-5b35423eaa83","Type":"ContainerStarted","Data":"c3ec812ab41e5290d980bb5e5141f4cbaa1788843d0710858dbf50ebfaf711fe"} Dec 03 12:44:29 crc kubenswrapper[4990]: I1203 12:44:29.192184 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" event={"ID":"f4b5755f-a0e8-4104-b4d1-5b35423eaa83","Type":"ContainerStarted","Data":"baedba0ade8c3b63574df8eb5d6dda8cb94a22b0edd8c5bcab8e57429003cb3d"} Dec 03 12:44:29 crc kubenswrapper[4990]: I1203 12:44:29.192313 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:33 crc kubenswrapper[4990]: I1203 12:44:33.287279 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:44:33 crc kubenswrapper[4990]: I1203 12:44:33.288218 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:44:33 crc kubenswrapper[4990]: I1203 12:44:33.288289 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:44:33 crc kubenswrapper[4990]: I1203 12:44:33.289168 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ff22e0abfe7a9119c84cea1563bea69a1449b575844b4f78fdcfa08778a9149"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 12:44:33 crc kubenswrapper[4990]: I1203 12:44:33.289244 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://8ff22e0abfe7a9119c84cea1563bea69a1449b575844b4f78fdcfa08778a9149" gracePeriod=600 Dec 03 12:44:34 crc kubenswrapper[4990]: I1203 12:44:34.232803 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="8ff22e0abfe7a9119c84cea1563bea69a1449b575844b4f78fdcfa08778a9149" exitCode=0 Dec 03 12:44:34 crc kubenswrapper[4990]: I1203 12:44:34.232895 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"8ff22e0abfe7a9119c84cea1563bea69a1449b575844b4f78fdcfa08778a9149"} Dec 03 12:44:34 crc kubenswrapper[4990]: I1203 12:44:34.233257 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"daf99345f123e10ced8b9ee63d4b06d27556fb355d285a58334782f921bd8551"} Dec 03 12:44:34 crc kubenswrapper[4990]: I1203 12:44:34.233290 4990 scope.go:117] "RemoveContainer" containerID="096e82c45b4b352e89d341a49b67638f723e9ba97670ebafa89a56f5ebdec156" Dec 03 12:44:34 crc kubenswrapper[4990]: I1203 12:44:34.260527 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" podStartSLOduration=7.260495174 podStartE2EDuration="7.260495174s" podCreationTimestamp="2025-12-03 12:44:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:29.230582142 +0000 UTC m=+417.372493381" watchObservedRunningTime="2025-12-03 12:44:34.260495174 +0000 UTC m=+422.402406413" Dec 03 12:44:35 crc kubenswrapper[4990]: I1203 12:44:35.527177 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kjsgh"] Dec 03 12:44:35 crc kubenswrapper[4990]: I1203 12:44:35.527875 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kjsgh" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="registry-server" containerID="cri-o://b3a06d33410d58cd038e690a65230a7969614cb6299d1357f88d8fe8ff737a0a" gracePeriod=2 Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.252350 4990 generic.go:334] "Generic (PLEG): container finished" podID="572c65be-7211-47d9-805e-081d07fbe844" containerID="b3a06d33410d58cd038e690a65230a7969614cb6299d1357f88d8fe8ff737a0a" exitCode=0 Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.252412 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerDied","Data":"b3a06d33410d58cd038e690a65230a7969614cb6299d1357f88d8fe8ff737a0a"} Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.556366 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.697157 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-utilities\") pod \"572c65be-7211-47d9-805e-081d07fbe844\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.697315 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5pqb\" (UniqueName: \"kubernetes.io/projected/572c65be-7211-47d9-805e-081d07fbe844-kube-api-access-v5pqb\") pod \"572c65be-7211-47d9-805e-081d07fbe844\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.697375 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-catalog-content\") pod \"572c65be-7211-47d9-805e-081d07fbe844\" (UID: \"572c65be-7211-47d9-805e-081d07fbe844\") " Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.698794 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-utilities" (OuterVolumeSpecName: "utilities") pod "572c65be-7211-47d9-805e-081d07fbe844" (UID: "572c65be-7211-47d9-805e-081d07fbe844"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.707351 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572c65be-7211-47d9-805e-081d07fbe844-kube-api-access-v5pqb" (OuterVolumeSpecName: "kube-api-access-v5pqb") pod "572c65be-7211-47d9-805e-081d07fbe844" (UID: "572c65be-7211-47d9-805e-081d07fbe844"). InnerVolumeSpecName "kube-api-access-v5pqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.762175 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "572c65be-7211-47d9-805e-081d07fbe844" (UID: "572c65be-7211-47d9-805e-081d07fbe844"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.799218 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.799272 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5pqb\" (UniqueName: \"kubernetes.io/projected/572c65be-7211-47d9-805e-081d07fbe844-kube-api-access-v5pqb\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:36 crc kubenswrapper[4990]: I1203 12:44:36.799289 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572c65be-7211-47d9-805e-081d07fbe844-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.262089 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjsgh" event={"ID":"572c65be-7211-47d9-805e-081d07fbe844","Type":"ContainerDied","Data":"be6dae5a15c07a0ec8396c205ab2c0a26f5d2fdf7425803e82d3b194e9a68e58"} Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.262180 4990 scope.go:117] "RemoveContainer" containerID="b3a06d33410d58cd038e690a65230a7969614cb6299d1357f88d8fe8ff737a0a" Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.262208 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjsgh" Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.282096 4990 scope.go:117] "RemoveContainer" containerID="9be9f927b0eb3489b7566fcd0bdb56ad3c9ea2b8bd01ffbcb38959c5d63ba73b" Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.319724 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kjsgh"] Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.324180 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kjsgh"] Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.327840 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tprnr"] Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.328103 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tprnr" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="registry-server" containerID="cri-o://ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" gracePeriod=2 Dec 03 12:44:37 crc kubenswrapper[4990]: I1203 12:44:37.330136 4990 scope.go:117] "RemoveContainer" containerID="d2c2ad962f16fbd9d4e23f30b5af1304d095931b50bcc1e86fbd5a5539af9056" Dec 03 12:44:37 crc kubenswrapper[4990]: E1203 12:44:37.416785 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod572c65be_7211_47d9_805e_081d07fbe844.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded381166_1062_4ca5_a553_81ab9681698e.slice/crio-conmon-ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod572c65be_7211_47d9_805e_081d07fbe844.slice/crio-be6dae5a15c07a0ec8396c205ab2c0a26f5d2fdf7425803e82d3b194e9a68e58\": RecentStats: unable to find data in memory cache]" Dec 03 12:44:37 crc kubenswrapper[4990]: E1203 12:44:37.571907 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe is running failed: container process not found" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 12:44:37 crc kubenswrapper[4990]: E1203 12:44:37.573167 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe is running failed: container process not found" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 12:44:37 crc kubenswrapper[4990]: E1203 12:44:37.573710 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe is running failed: container process not found" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 12:44:37 crc kubenswrapper[4990]: E1203 12:44:37.573752 4990 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-tprnr" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="registry-server" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.073525 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.130310 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbxg8"] Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.130966 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cbxg8" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="registry-server" containerID="cri-o://8d6079685da74fad7c22d1fbec9328b4600943f35c4084c3b70e30aa3cc66ddf" gracePeriod=2 Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.217990 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-utilities\") pod \"ed381166-1062-4ca5-a553-81ab9681698e\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.218117 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-catalog-content\") pod \"ed381166-1062-4ca5-a553-81ab9681698e\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.218275 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc8t9\" (UniqueName: \"kubernetes.io/projected/ed381166-1062-4ca5-a553-81ab9681698e-kube-api-access-cc8t9\") pod \"ed381166-1062-4ca5-a553-81ab9681698e\" (UID: \"ed381166-1062-4ca5-a553-81ab9681698e\") " Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.219817 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-utilities" (OuterVolumeSpecName: "utilities") pod "ed381166-1062-4ca5-a553-81ab9681698e" (UID: "ed381166-1062-4ca5-a553-81ab9681698e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.226525 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed381166-1062-4ca5-a553-81ab9681698e-kube-api-access-cc8t9" (OuterVolumeSpecName: "kube-api-access-cc8t9") pod "ed381166-1062-4ca5-a553-81ab9681698e" (UID: "ed381166-1062-4ca5-a553-81ab9681698e"). InnerVolumeSpecName "kube-api-access-cc8t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.273374 4990 generic.go:334] "Generic (PLEG): container finished" podID="ed381166-1062-4ca5-a553-81ab9681698e" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" exitCode=0 Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.273725 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tprnr" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.274285 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572c65be-7211-47d9-805e-081d07fbe844" path="/var/lib/kubelet/pods/572c65be-7211-47d9-805e-081d07fbe844/volumes" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.281289 4990 generic.go:334] "Generic (PLEG): container finished" podID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerID="8d6079685da74fad7c22d1fbec9328b4600943f35c4084c3b70e30aa3cc66ddf" exitCode=0 Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.284013 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerDied","Data":"ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe"} Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.284093 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tprnr" event={"ID":"ed381166-1062-4ca5-a553-81ab9681698e","Type":"ContainerDied","Data":"cef556ba78ee1616197a0dc73025df70cc0a2ccfa42b4613d52a025795f2b597"} Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.284114 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbxg8" event={"ID":"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030","Type":"ContainerDied","Data":"8d6079685da74fad7c22d1fbec9328b4600943f35c4084c3b70e30aa3cc66ddf"} Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.284132 4990 scope.go:117] "RemoveContainer" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.288210 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed381166-1062-4ca5-a553-81ab9681698e" (UID: "ed381166-1062-4ca5-a553-81ab9681698e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.306509 4990 scope.go:117] "RemoveContainer" containerID="c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.320807 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.320873 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed381166-1062-4ca5-a553-81ab9681698e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.320891 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc8t9\" (UniqueName: \"kubernetes.io/projected/ed381166-1062-4ca5-a553-81ab9681698e-kube-api-access-cc8t9\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.327400 4990 scope.go:117] "RemoveContainer" containerID="ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.348655 4990 scope.go:117] "RemoveContainer" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" Dec 03 12:44:38 crc kubenswrapper[4990]: E1203 12:44:38.350595 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe\": container with ID starting with ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe not found: ID does not exist" containerID="ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.350679 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe"} err="failed to get container status \"ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe\": rpc error: code = NotFound desc = could not find container \"ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe\": container with ID starting with ecba233b4f0437e9a903b6c8ced4f25f285ae176f2c75269e4dd3877a2c6adfe not found: ID does not exist" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.350738 4990 scope.go:117] "RemoveContainer" containerID="c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589" Dec 03 12:44:38 crc kubenswrapper[4990]: E1203 12:44:38.351260 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589\": container with ID starting with c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589 not found: ID does not exist" containerID="c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.351310 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589"} err="failed to get container status \"c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589\": rpc error: code = NotFound desc = could not find container \"c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589\": container with ID starting with c7eb2545d0b601537a36300673cf9edc30cc31bf779d938301b0382b91e2b589 not found: ID does not exist" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.351346 4990 scope.go:117] "RemoveContainer" containerID="ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175" Dec 03 12:44:38 crc kubenswrapper[4990]: E1203 12:44:38.351939 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175\": container with ID starting with ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175 not found: ID does not exist" containerID="ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.351986 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175"} err="failed to get container status \"ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175\": rpc error: code = NotFound desc = could not find container \"ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175\": container with ID starting with ca75d7858272dc0dfcd03d718ec756ce6d7efca4975566d5d8fdc6ed9156f175 not found: ID does not exist" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.634410 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tprnr"] Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.639303 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tprnr"] Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.650961 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.731774 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-utilities\") pod \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.732126 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd42h\" (UniqueName: \"kubernetes.io/projected/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-kube-api-access-jd42h\") pod \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.732304 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-catalog-content\") pod \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\" (UID: \"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030\") " Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.733029 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-utilities" (OuterVolumeSpecName: "utilities") pod "33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" (UID: "33b6b642-d3d4-4a33-9bf0-25d5c2aa9030"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.734497 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.736033 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-kube-api-access-jd42h" (OuterVolumeSpecName: "kube-api-access-jd42h") pod "33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" (UID: "33b6b642-d3d4-4a33-9bf0-25d5c2aa9030"). InnerVolumeSpecName "kube-api-access-jd42h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.751357 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" (UID: "33b6b642-d3d4-4a33-9bf0-25d5c2aa9030"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.835673 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd42h\" (UniqueName: \"kubernetes.io/projected/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-kube-api-access-jd42h\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:38 crc kubenswrapper[4990]: I1203 12:44:38.835726 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.292746 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbxg8" event={"ID":"33b6b642-d3d4-4a33-9bf0-25d5c2aa9030","Type":"ContainerDied","Data":"70eeb3b465b52601a826e86e0e281394dfa0bce3dca6ddcc2aca1e580685d292"} Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.293263 4990 scope.go:117] "RemoveContainer" containerID="8d6079685da74fad7c22d1fbec9328b4600943f35c4084c3b70e30aa3cc66ddf" Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.292821 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbxg8" Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.328674 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbxg8"] Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.333353 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbxg8"] Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.339246 4990 scope.go:117] "RemoveContainer" containerID="8d1cddaff8fc11f602e6bba7598aa77a3eb2bb33e5d5f233160020c224883073" Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.360468 4990 scope.go:117] "RemoveContainer" containerID="dd143e7dfee45916b619826274787c7470faa2d2367ef9e0b0c0c17c776ef2d8" Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.421690 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-m4xsj"] Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.421969 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" podUID="9f57f24a-6040-41ff-9077-75038fc76de4" containerName="controller-manager" containerID="cri-o://330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2" gracePeriod=30 Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.450476 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq"] Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.450835 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" podUID="9c20895f-e512-47fe-b62b-d7d47080bbd1" containerName="route-controller-manager" containerID="cri-o://7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70" gracePeriod=30 Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.582286 4990 patch_prober.go:28] interesting pod/controller-manager-55b446ff67-m4xsj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body= Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.582355 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" podUID="9f57f24a-6040-41ff-9077-75038fc76de4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.927937 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sfvnj"] Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.928261 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sfvnj" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="registry-server" containerID="cri-o://a594425cfa5eb664e8cb7d9236624bfd80eb31ec1f7c0c254b9e9902a0fc09e3" gracePeriod=2 Dec 03 12:44:39 crc kubenswrapper[4990]: I1203 12:44:39.980031 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.154227 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrwck\" (UniqueName: \"kubernetes.io/projected/9c20895f-e512-47fe-b62b-d7d47080bbd1-kube-api-access-qrwck\") pod \"9c20895f-e512-47fe-b62b-d7d47080bbd1\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.154289 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-config\") pod \"9c20895f-e512-47fe-b62b-d7d47080bbd1\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.154345 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-client-ca\") pod \"9c20895f-e512-47fe-b62b-d7d47080bbd1\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.154415 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c20895f-e512-47fe-b62b-d7d47080bbd1-serving-cert\") pod \"9c20895f-e512-47fe-b62b-d7d47080bbd1\" (UID: \"9c20895f-e512-47fe-b62b-d7d47080bbd1\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.155912 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-config" (OuterVolumeSpecName: "config") pod "9c20895f-e512-47fe-b62b-d7d47080bbd1" (UID: "9c20895f-e512-47fe-b62b-d7d47080bbd1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.155900 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-client-ca" (OuterVolumeSpecName: "client-ca") pod "9c20895f-e512-47fe-b62b-d7d47080bbd1" (UID: "9c20895f-e512-47fe-b62b-d7d47080bbd1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.176869 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c20895f-e512-47fe-b62b-d7d47080bbd1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9c20895f-e512-47fe-b62b-d7d47080bbd1" (UID: "9c20895f-e512-47fe-b62b-d7d47080bbd1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.176982 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c20895f-e512-47fe-b62b-d7d47080bbd1-kube-api-access-qrwck" (OuterVolumeSpecName: "kube-api-access-qrwck") pod "9c20895f-e512-47fe-b62b-d7d47080bbd1" (UID: "9c20895f-e512-47fe-b62b-d7d47080bbd1"). InnerVolumeSpecName "kube-api-access-qrwck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.187311 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.256824 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c20895f-e512-47fe-b62b-d7d47080bbd1-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.256865 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrwck\" (UniqueName: \"kubernetes.io/projected/9c20895f-e512-47fe-b62b-d7d47080bbd1-kube-api-access-qrwck\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.256877 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.256892 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c20895f-e512-47fe-b62b-d7d47080bbd1-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.274145 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" path="/var/lib/kubelet/pods/33b6b642-d3d4-4a33-9bf0-25d5c2aa9030/volumes" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.275095 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed381166-1062-4ca5-a553-81ab9681698e" path="/var/lib/kubelet/pods/ed381166-1062-4ca5-a553-81ab9681698e/volumes" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.306742 4990 generic.go:334] "Generic (PLEG): container finished" podID="9c20895f-e512-47fe-b62b-d7d47080bbd1" containerID="7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70" exitCode=0 Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.306829 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" event={"ID":"9c20895f-e512-47fe-b62b-d7d47080bbd1","Type":"ContainerDied","Data":"7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70"} Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.306858 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" event={"ID":"9c20895f-e512-47fe-b62b-d7d47080bbd1","Type":"ContainerDied","Data":"716b7c4617c87694fb013050f2d5ea470a35d44584d547c22cc90e8fcda1ac08"} Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.306868 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.306874 4990 scope.go:117] "RemoveContainer" containerID="7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.316423 4990 generic.go:334] "Generic (PLEG): container finished" podID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerID="a594425cfa5eb664e8cb7d9236624bfd80eb31ec1f7c0c254b9e9902a0fc09e3" exitCode=0 Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.316519 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerDied","Data":"a594425cfa5eb664e8cb7d9236624bfd80eb31ec1f7c0c254b9e9902a0fc09e3"} Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.320029 4990 generic.go:334] "Generic (PLEG): container finished" podID="9f57f24a-6040-41ff-9077-75038fc76de4" containerID="330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2" exitCode=0 Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.320063 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" event={"ID":"9f57f24a-6040-41ff-9077-75038fc76de4","Type":"ContainerDied","Data":"330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2"} Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.320084 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" event={"ID":"9f57f24a-6040-41ff-9077-75038fc76de4","Type":"ContainerDied","Data":"c2087b83781f8ebaae246dba664a038e78f3339fc2233a82eb02b8395d6b976a"} Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.320141 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b446ff67-m4xsj" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.334056 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq"] Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.341370 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646bddd5f-drhqq"] Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.357717 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-config\") pod \"9f57f24a-6040-41ff-9077-75038fc76de4\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.357774 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-proxy-ca-bundles\") pod \"9f57f24a-6040-41ff-9077-75038fc76de4\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.357851 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wdp6\" (UniqueName: \"kubernetes.io/projected/9f57f24a-6040-41ff-9077-75038fc76de4-kube-api-access-6wdp6\") pod \"9f57f24a-6040-41ff-9077-75038fc76de4\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.357873 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f57f24a-6040-41ff-9077-75038fc76de4-serving-cert\") pod \"9f57f24a-6040-41ff-9077-75038fc76de4\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.357902 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-client-ca\") pod \"9f57f24a-6040-41ff-9077-75038fc76de4\" (UID: \"9f57f24a-6040-41ff-9077-75038fc76de4\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.359032 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-client-ca" (OuterVolumeSpecName: "client-ca") pod "9f57f24a-6040-41ff-9077-75038fc76de4" (UID: "9f57f24a-6040-41ff-9077-75038fc76de4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.359154 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9f57f24a-6040-41ff-9077-75038fc76de4" (UID: "9f57f24a-6040-41ff-9077-75038fc76de4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.359813 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-config" (OuterVolumeSpecName: "config") pod "9f57f24a-6040-41ff-9077-75038fc76de4" (UID: "9f57f24a-6040-41ff-9077-75038fc76de4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.362837 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f57f24a-6040-41ff-9077-75038fc76de4-kube-api-access-6wdp6" (OuterVolumeSpecName: "kube-api-access-6wdp6") pod "9f57f24a-6040-41ff-9077-75038fc76de4" (UID: "9f57f24a-6040-41ff-9077-75038fc76de4"). InnerVolumeSpecName "kube-api-access-6wdp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.363005 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f57f24a-6040-41ff-9077-75038fc76de4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9f57f24a-6040-41ff-9077-75038fc76de4" (UID: "9f57f24a-6040-41ff-9077-75038fc76de4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.365274 4990 scope.go:117] "RemoveContainer" containerID="7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70" Dec 03 12:44:40 crc kubenswrapper[4990]: E1203 12:44:40.370008 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70\": container with ID starting with 7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70 not found: ID does not exist" containerID="7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.370081 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70"} err="failed to get container status \"7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70\": rpc error: code = NotFound desc = could not find container \"7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70\": container with ID starting with 7aa9945846316f22ea09b018f437ee15645c9ee340e7c5fcf9ade0b512710d70 not found: ID does not exist" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.370126 4990 scope.go:117] "RemoveContainer" containerID="330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.372984 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.409890 4990 scope.go:117] "RemoveContainer" containerID="330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2" Dec 03 12:44:40 crc kubenswrapper[4990]: E1203 12:44:40.410755 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2\": container with ID starting with 330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2 not found: ID does not exist" containerID="330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.410809 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2"} err="failed to get container status \"330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2\": rpc error: code = NotFound desc = could not find container \"330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2\": container with ID starting with 330e1383e0dd2125a40122a8519f830d529a9363e744152eb0e41c8cabe008c2 not found: ID does not exist" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.459168 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tthkk\" (UniqueName: \"kubernetes.io/projected/2069abff-1d0a-42a0-8929-75acceecbe8b-kube-api-access-tthkk\") pod \"2069abff-1d0a-42a0-8929-75acceecbe8b\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460334 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-catalog-content\") pod \"2069abff-1d0a-42a0-8929-75acceecbe8b\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460370 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-utilities\") pod \"2069abff-1d0a-42a0-8929-75acceecbe8b\" (UID: \"2069abff-1d0a-42a0-8929-75acceecbe8b\") " Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460752 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460771 4990 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460788 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wdp6\" (UniqueName: \"kubernetes.io/projected/9f57f24a-6040-41ff-9077-75038fc76de4-kube-api-access-6wdp6\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460801 4990 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f57f24a-6040-41ff-9077-75038fc76de4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.460810 4990 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f57f24a-6040-41ff-9077-75038fc76de4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.461578 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-utilities" (OuterVolumeSpecName: "utilities") pod "2069abff-1d0a-42a0-8929-75acceecbe8b" (UID: "2069abff-1d0a-42a0-8929-75acceecbe8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.463903 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2069abff-1d0a-42a0-8929-75acceecbe8b-kube-api-access-tthkk" (OuterVolumeSpecName: "kube-api-access-tthkk") pod "2069abff-1d0a-42a0-8929-75acceecbe8b" (UID: "2069abff-1d0a-42a0-8929-75acceecbe8b"). InnerVolumeSpecName "kube-api-access-tthkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.561954 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tthkk\" (UniqueName: \"kubernetes.io/projected/2069abff-1d0a-42a0-8929-75acceecbe8b-kube-api-access-tthkk\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.562295 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.590362 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2069abff-1d0a-42a0-8929-75acceecbe8b" (UID: "2069abff-1d0a-42a0-8929-75acceecbe8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.656310 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-m4xsj"] Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.663652 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2069abff-1d0a-42a0-8929-75acceecbe8b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:40 crc kubenswrapper[4990]: I1203 12:44:40.664278 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-55b446ff67-m4xsj"] Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278062 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk"] Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278441 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278490 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278507 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278519 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278538 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278547 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278560 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278568 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278575 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278581 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278595 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278601 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278612 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278618 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278631 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278637 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278648 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278654 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="extract-content" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278667 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c20895f-e512-47fe-b62b-d7d47080bbd1" containerName="route-controller-manager" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278673 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c20895f-e512-47fe-b62b-d7d47080bbd1" containerName="route-controller-manager" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278679 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278685 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278693 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f57f24a-6040-41ff-9077-75038fc76de4" containerName="controller-manager" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278698 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f57f24a-6040-41ff-9077-75038fc76de4" containerName="controller-manager" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278708 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278715 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="extract-utilities" Dec 03 12:44:41 crc kubenswrapper[4990]: E1203 12:44:41.278725 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278738 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278867 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f57f24a-6040-41ff-9077-75038fc76de4" containerName="controller-manager" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278886 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="572c65be-7211-47d9-805e-081d07fbe844" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278896 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed381166-1062-4ca5-a553-81ab9681698e" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278931 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278940 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b6b642-d3d4-4a33-9bf0-25d5c2aa9030" containerName="registry-server" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.278949 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c20895f-e512-47fe-b62b-d7d47080bbd1" containerName="route-controller-manager" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.280055 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.281798 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66f46dd896-2gzgb"] Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.282675 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.286800 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.288168 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.290420 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.290645 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.290955 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.291509 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.291667 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.291717 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.291847 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.292084 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.292226 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.299642 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.302023 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk"] Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.302334 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.321707 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f46dd896-2gzgb"] Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.350822 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvnj" event={"ID":"2069abff-1d0a-42a0-8929-75acceecbe8b","Type":"ContainerDied","Data":"f873bd3fb51af49d95eb0e04d9ecd566d18fbb4a4561c303ec40ba6ae74776b2"} Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.350891 4990 scope.go:117] "RemoveContainer" containerID="a594425cfa5eb664e8cb7d9236624bfd80eb31ec1f7c0c254b9e9902a0fc09e3" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.350990 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvnj" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.371771 4990 scope.go:117] "RemoveContainer" containerID="4542dd6aa8c3f3b41dc07b034ae5e13683ffb8df40a97d28633fcee6e857b984" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374280 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c35b2d1a-9720-4b37-9930-d04e9453dd3b-serving-cert\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374339 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-config\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374383 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-serving-cert\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374409 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq8gp\" (UniqueName: \"kubernetes.io/projected/c35b2d1a-9720-4b37-9930-d04e9453dd3b-kube-api-access-xq8gp\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374469 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-client-ca\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374501 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-proxy-ca-bundles\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374531 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msgxr\" (UniqueName: \"kubernetes.io/projected/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-kube-api-access-msgxr\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374565 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-config\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.374604 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-client-ca\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.394551 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sfvnj"] Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.395375 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sfvnj"] Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.416071 4990 scope.go:117] "RemoveContainer" containerID="9935e4025e2531719e679337bceba3a8cc735f11fe2f938f2da6c76a4ac75962" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476048 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msgxr\" (UniqueName: \"kubernetes.io/projected/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-kube-api-access-msgxr\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476147 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-config\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476217 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-client-ca\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476246 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c35b2d1a-9720-4b37-9930-d04e9453dd3b-serving-cert\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476283 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-config\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476333 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-serving-cert\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476364 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq8gp\" (UniqueName: \"kubernetes.io/projected/c35b2d1a-9720-4b37-9930-d04e9453dd3b-kube-api-access-xq8gp\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476404 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-client-ca\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.476442 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-proxy-ca-bundles\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.477643 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-client-ca\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.477994 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-config\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.478317 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-client-ca\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.478708 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-proxy-ca-bundles\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.479109 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c35b2d1a-9720-4b37-9930-d04e9453dd3b-config\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.481048 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c35b2d1a-9720-4b37-9930-d04e9453dd3b-serving-cert\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.483189 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-serving-cert\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.499334 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msgxr\" (UniqueName: \"kubernetes.io/projected/c3a1a881-e718-4bbb-8cb5-1b4c71d47a19-kube-api-access-msgxr\") pod \"route-controller-manager-c7df7995b-665zk\" (UID: \"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19\") " pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.500488 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq8gp\" (UniqueName: \"kubernetes.io/projected/c35b2d1a-9720-4b37-9930-d04e9453dd3b-kube-api-access-xq8gp\") pod \"controller-manager-66f46dd896-2gzgb\" (UID: \"c35b2d1a-9720-4b37-9930-d04e9453dd3b\") " pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.600350 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.613432 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:41 crc kubenswrapper[4990]: I1203 12:44:41.948353 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f46dd896-2gzgb"] Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.110238 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk"] Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.272415 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2069abff-1d0a-42a0-8929-75acceecbe8b" path="/var/lib/kubelet/pods/2069abff-1d0a-42a0-8929-75acceecbe8b/volumes" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.274064 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c20895f-e512-47fe-b62b-d7d47080bbd1" path="/var/lib/kubelet/pods/9c20895f-e512-47fe-b62b-d7d47080bbd1/volumes" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.274726 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f57f24a-6040-41ff-9077-75038fc76de4" path="/var/lib/kubelet/pods/9f57f24a-6040-41ff-9077-75038fc76de4/volumes" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.358970 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" event={"ID":"c35b2d1a-9720-4b37-9930-d04e9453dd3b","Type":"ContainerStarted","Data":"49de15cb7080c4e5e5dffb3be6c0880a27ae28e70b10d9af81a6c15c50ea0340"} Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.359028 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" event={"ID":"c35b2d1a-9720-4b37-9930-d04e9453dd3b","Type":"ContainerStarted","Data":"b2661d79bedbb54e101f5ba6d993bf4a67ba6c85fac31bdb72528f23d065417d"} Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.359250 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.361128 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" event={"ID":"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19","Type":"ContainerStarted","Data":"b22a0836f943d384a5afd91fcecfc85288002189b4b3f4e367482c63ca004649"} Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.361172 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" event={"ID":"c3a1a881-e718-4bbb-8cb5-1b4c71d47a19","Type":"ContainerStarted","Data":"0b3b83b8eeb552b6b4ffa7317d8c8d3dadce12545b17a1620ddacccc389fe0a7"} Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.361258 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.363184 4990 patch_prober.go:28] interesting pod/route-controller-manager-c7df7995b-665zk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": dial tcp 10.217.0.63:8443: connect: connection refused" start-of-body= Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.363244 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" podUID="c3a1a881-e718-4bbb-8cb5-1b4c71d47a19" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": dial tcp 10.217.0.63:8443: connect: connection refused" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.368155 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.416294 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" podStartSLOduration=3.416263126 podStartE2EDuration="3.416263126s" podCreationTimestamp="2025-12-03 12:44:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:42.412846785 +0000 UTC m=+430.554758014" watchObservedRunningTime="2025-12-03 12:44:42.416263126 +0000 UTC m=+430.558174355" Dec 03 12:44:42 crc kubenswrapper[4990]: I1203 12:44:42.417263 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66f46dd896-2gzgb" podStartSLOduration=3.4172524810000002 podStartE2EDuration="3.417252481s" podCreationTimestamp="2025-12-03 12:44:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:42.387778081 +0000 UTC m=+430.529689310" watchObservedRunningTime="2025-12-03 12:44:42.417252481 +0000 UTC m=+430.559163710" Dec 03 12:44:43 crc kubenswrapper[4990]: I1203 12:44:43.373844 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c7df7995b-665zk" Dec 03 12:44:47 crc kubenswrapper[4990]: I1203 12:44:47.736974 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9hdr5" Dec 03 12:44:47 crc kubenswrapper[4990]: I1203 12:44:47.807847 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-twmj9"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.649802 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jxn9m"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.651012 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jxn9m" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="registry-server" containerID="cri-o://02d693cb3abf43b38366de04886fea8d8fcf8aa59103b1ac5530b5cd332785e6" gracePeriod=30 Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.662546 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j98rl"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.662966 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j98rl" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="registry-server" containerID="cri-o://8f42b33dc07af86a83a56129107ca063742b5b604584ab5cfb4584c16ef8fc3a" gracePeriod=30 Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.671270 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gmsz9"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.674221 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" containerID="cri-o://cdbd7460ad4601847233699deb0e5f3d3d63e279d3c37733744f03a1406cf537" gracePeriod=30 Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.686038 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbqqz"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.686320 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dbqqz" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="registry-server" containerID="cri-o://98ce943cfb69275a66c986eafff99bc65a3a80144327c59fb8fa0c49f9df9f23" gracePeriod=30 Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.698343 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zn87g"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.698671 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zn87g" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="registry-server" containerID="cri-o://abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0" gracePeriod=30 Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.708884 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9n6hl"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.709850 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.730200 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9n6hl"] Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.842951 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a0b82ca-6250-4ca2-abb5-2704907e4f98-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.843013 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kntq\" (UniqueName: \"kubernetes.io/projected/8a0b82ca-6250-4ca2-abb5-2704907e4f98-kube-api-access-7kntq\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.843107 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8a0b82ca-6250-4ca2-abb5-2704907e4f98-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.944196 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a0b82ca-6250-4ca2-abb5-2704907e4f98-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.944254 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kntq\" (UniqueName: \"kubernetes.io/projected/8a0b82ca-6250-4ca2-abb5-2704907e4f98-kube-api-access-7kntq\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.944324 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8a0b82ca-6250-4ca2-abb5-2704907e4f98-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.948481 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a0b82ca-6250-4ca2-abb5-2704907e4f98-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.959872 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8a0b82ca-6250-4ca2-abb5-2704907e4f98-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:49 crc kubenswrapper[4990]: I1203 12:44:49.969288 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kntq\" (UniqueName: \"kubernetes.io/projected/8a0b82ca-6250-4ca2-abb5-2704907e4f98-kube-api-access-7kntq\") pod \"marketplace-operator-79b997595-9n6hl\" (UID: \"8a0b82ca-6250-4ca2-abb5-2704907e4f98\") " pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.046118 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:50 crc kubenswrapper[4990]: E1203 12:44:50.200043 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0 is running failed: container process not found" containerID="abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 12:44:50 crc kubenswrapper[4990]: E1203 12:44:50.201213 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0 is running failed: container process not found" containerID="abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 12:44:50 crc kubenswrapper[4990]: E1203 12:44:50.202314 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0 is running failed: container process not found" containerID="abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 12:44:50 crc kubenswrapper[4990]: E1203 12:44:50.202355 4990 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-zn87g" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="registry-server" Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.416076 4990 generic.go:334] "Generic (PLEG): container finished" podID="89cb569a-c36f-481c-b5a0-00694046c00c" containerID="8f42b33dc07af86a83a56129107ca063742b5b604584ab5cfb4584c16ef8fc3a" exitCode=0 Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.416155 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerDied","Data":"8f42b33dc07af86a83a56129107ca063742b5b604584ab5cfb4584c16ef8fc3a"} Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.434549 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gmsz9_2870df3c-124c-4a2d-8f88-4fec5f65740c/marketplace-operator/1.log" Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.434612 4990 generic.go:334] "Generic (PLEG): container finished" podID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerID="cdbd7460ad4601847233699deb0e5f3d3d63e279d3c37733744f03a1406cf537" exitCode=0 Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.434757 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerDied","Data":"cdbd7460ad4601847233699deb0e5f3d3d63e279d3c37733744f03a1406cf537"} Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.434821 4990 scope.go:117] "RemoveContainer" containerID="caadcf267bb9499dd5a413687a3708feda6c22ed87ce1f1f4706f019af7fec70" Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.448259 4990 generic.go:334] "Generic (PLEG): container finished" podID="8108de1d-75ea-419d-9658-e5a035088b36" containerID="02d693cb3abf43b38366de04886fea8d8fcf8aa59103b1ac5530b5cd332785e6" exitCode=0 Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.448363 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerDied","Data":"02d693cb3abf43b38366de04886fea8d8fcf8aa59103b1ac5530b5cd332785e6"} Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.456844 4990 generic.go:334] "Generic (PLEG): container finished" podID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerID="98ce943cfb69275a66c986eafff99bc65a3a80144327c59fb8fa0c49f9df9f23" exitCode=0 Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.456948 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbqqz" event={"ID":"ddec4b1a-b34e-45f7-95e6-a04d29dde82c","Type":"ContainerDied","Data":"98ce943cfb69275a66c986eafff99bc65a3a80144327c59fb8fa0c49f9df9f23"} Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.461874 4990 generic.go:334] "Generic (PLEG): container finished" podID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerID="abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0" exitCode=0 Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.461965 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerDied","Data":"abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0"} Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.514250 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9n6hl"] Dec 03 12:44:50 crc kubenswrapper[4990]: I1203 12:44:50.919918 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.062090 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.067746 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.069102 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-utilities\") pod \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.069175 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zj55\" (UniqueName: \"kubernetes.io/projected/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-kube-api-access-5zj55\") pod \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.069239 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-catalog-content\") pod \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\" (UID: \"ddec4b1a-b34e-45f7-95e6-a04d29dde82c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.070686 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-utilities" (OuterVolumeSpecName: "utilities") pod "ddec4b1a-b34e-45f7-95e6-a04d29dde82c" (UID: "ddec4b1a-b34e-45f7-95e6-a04d29dde82c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.076415 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-kube-api-access-5zj55" (OuterVolumeSpecName: "kube-api-access-5zj55") pod "ddec4b1a-b34e-45f7-95e6-a04d29dde82c" (UID: "ddec4b1a-b34e-45f7-95e6-a04d29dde82c"). InnerVolumeSpecName "kube-api-access-5zj55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.077580 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.086961 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.095260 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddec4b1a-b34e-45f7-95e6-a04d29dde82c" (UID: "ddec4b1a-b34e-45f7-95e6-a04d29dde82c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.170918 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-operator-metrics\") pod \"2870df3c-124c-4a2d-8f88-4fec5f65740c\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.173971 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-catalog-content\") pod \"8108de1d-75ea-419d-9658-e5a035088b36\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174493 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-utilities\") pod \"89cb569a-c36f-481c-b5a0-00694046c00c\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174552 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nrdf\" (UniqueName: \"kubernetes.io/projected/34bb2446-2a47-4b35-b5a1-13d8de74822e-kube-api-access-5nrdf\") pod \"34bb2446-2a47-4b35-b5a1-13d8de74822e\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174647 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-catalog-content\") pod \"34bb2446-2a47-4b35-b5a1-13d8de74822e\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174672 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-466jz\" (UniqueName: \"kubernetes.io/projected/89cb569a-c36f-481c-b5a0-00694046c00c-kube-api-access-466jz\") pod \"89cb569a-c36f-481c-b5a0-00694046c00c\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174725 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jnqm\" (UniqueName: \"kubernetes.io/projected/8108de1d-75ea-419d-9658-e5a035088b36-kube-api-access-8jnqm\") pod \"8108de1d-75ea-419d-9658-e5a035088b36\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174756 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-trusted-ca\") pod \"2870df3c-124c-4a2d-8f88-4fec5f65740c\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174811 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-catalog-content\") pod \"89cb569a-c36f-481c-b5a0-00694046c00c\" (UID: \"89cb569a-c36f-481c-b5a0-00694046c00c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174877 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-utilities\") pod \"34bb2446-2a47-4b35-b5a1-13d8de74822e\" (UID: \"34bb2446-2a47-4b35-b5a1-13d8de74822e\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.174961 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq4f9\" (UniqueName: \"kubernetes.io/projected/2870df3c-124c-4a2d-8f88-4fec5f65740c-kube-api-access-vq4f9\") pod \"2870df3c-124c-4a2d-8f88-4fec5f65740c\" (UID: \"2870df3c-124c-4a2d-8f88-4fec5f65740c\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.175037 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-utilities\") pod \"8108de1d-75ea-419d-9658-e5a035088b36\" (UID: \"8108de1d-75ea-419d-9658-e5a035088b36\") " Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.176165 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.176223 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zj55\" (UniqueName: \"kubernetes.io/projected/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-kube-api-access-5zj55\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.176238 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddec4b1a-b34e-45f7-95e6-a04d29dde82c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.178143 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-utilities" (OuterVolumeSpecName: "utilities") pod "8108de1d-75ea-419d-9658-e5a035088b36" (UID: "8108de1d-75ea-419d-9658-e5a035088b36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.179699 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-utilities" (OuterVolumeSpecName: "utilities") pod "34bb2446-2a47-4b35-b5a1-13d8de74822e" (UID: "34bb2446-2a47-4b35-b5a1-13d8de74822e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.182601 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2870df3c-124c-4a2d-8f88-4fec5f65740c" (UID: "2870df3c-124c-4a2d-8f88-4fec5f65740c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.183857 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2870df3c-124c-4a2d-8f88-4fec5f65740c" (UID: "2870df3c-124c-4a2d-8f88-4fec5f65740c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.185515 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89cb569a-c36f-481c-b5a0-00694046c00c-kube-api-access-466jz" (OuterVolumeSpecName: "kube-api-access-466jz") pod "89cb569a-c36f-481c-b5a0-00694046c00c" (UID: "89cb569a-c36f-481c-b5a0-00694046c00c"). InnerVolumeSpecName "kube-api-access-466jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.186414 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-utilities" (OuterVolumeSpecName: "utilities") pod "89cb569a-c36f-481c-b5a0-00694046c00c" (UID: "89cb569a-c36f-481c-b5a0-00694046c00c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.195720 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34bb2446-2a47-4b35-b5a1-13d8de74822e-kube-api-access-5nrdf" (OuterVolumeSpecName: "kube-api-access-5nrdf") pod "34bb2446-2a47-4b35-b5a1-13d8de74822e" (UID: "34bb2446-2a47-4b35-b5a1-13d8de74822e"). InnerVolumeSpecName "kube-api-access-5nrdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.195840 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8108de1d-75ea-419d-9658-e5a035088b36-kube-api-access-8jnqm" (OuterVolumeSpecName: "kube-api-access-8jnqm") pod "8108de1d-75ea-419d-9658-e5a035088b36" (UID: "8108de1d-75ea-419d-9658-e5a035088b36"). InnerVolumeSpecName "kube-api-access-8jnqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.196661 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2870df3c-124c-4a2d-8f88-4fec5f65740c-kube-api-access-vq4f9" (OuterVolumeSpecName: "kube-api-access-vq4f9") pod "2870df3c-124c-4a2d-8f88-4fec5f65740c" (UID: "2870df3c-124c-4a2d-8f88-4fec5f65740c"). InnerVolumeSpecName "kube-api-access-vq4f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.266610 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89cb569a-c36f-481c-b5a0-00694046c00c" (UID: "89cb569a-c36f-481c-b5a0-00694046c00c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277550 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277605 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nrdf\" (UniqueName: \"kubernetes.io/projected/34bb2446-2a47-4b35-b5a1-13d8de74822e-kube-api-access-5nrdf\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277618 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-466jz\" (UniqueName: \"kubernetes.io/projected/89cb569a-c36f-481c-b5a0-00694046c00c-kube-api-access-466jz\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277630 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jnqm\" (UniqueName: \"kubernetes.io/projected/8108de1d-75ea-419d-9658-e5a035088b36-kube-api-access-8jnqm\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277642 4990 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277652 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277664 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cb569a-c36f-481c-b5a0-00694046c00c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277676 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq4f9\" (UniqueName: \"kubernetes.io/projected/2870df3c-124c-4a2d-8f88-4fec5f65740c-kube-api-access-vq4f9\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277688 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.277698 4990 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2870df3c-124c-4a2d-8f88-4fec5f65740c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.284749 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8108de1d-75ea-419d-9658-e5a035088b36" (UID: "8108de1d-75ea-419d-9658-e5a035088b36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.329897 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34bb2446-2a47-4b35-b5a1-13d8de74822e" (UID: "34bb2446-2a47-4b35-b5a1-13d8de74822e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.378824 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8108de1d-75ea-419d-9658-e5a035088b36-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.378875 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34bb2446-2a47-4b35-b5a1-13d8de74822e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.472917 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" event={"ID":"2870df3c-124c-4a2d-8f88-4fec5f65740c","Type":"ContainerDied","Data":"ec756ac85c90651c5bab3988b6f4ea7975dee2e4fa0a3807209cd2cbce049a7d"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.472972 4990 scope.go:117] "RemoveContainer" containerID="cdbd7460ad4601847233699deb0e5f3d3d63e279d3c37733744f03a1406cf537" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.472969 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gmsz9" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.477212 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxn9m" event={"ID":"8108de1d-75ea-419d-9658-e5a035088b36","Type":"ContainerDied","Data":"26d08bb85944788450f37a8b4d937de0db7a6b94fde9e0ce0ee3f3bd56961d28"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.477248 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxn9m" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.480175 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zn87g" event={"ID":"34bb2446-2a47-4b35-b5a1-13d8de74822e","Type":"ContainerDied","Data":"4ccfd97112ed203840a201232e8b26b75883eca509b285caadab3e26606a8dbf"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.480259 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zn87g" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.484180 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbqqz" event={"ID":"ddec4b1a-b34e-45f7-95e6-a04d29dde82c","Type":"ContainerDied","Data":"7a83fc586282919f5e05615f3cc72d05a6d738424fe93e0be8f8107656db5e63"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.484322 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbqqz" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.487014 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j98rl" event={"ID":"89cb569a-c36f-481c-b5a0-00694046c00c","Type":"ContainerDied","Data":"d8e3cd47f79a146ff14c7e46b1b5a45ffba08627a0aecb3a16e97cc100a5411d"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.487046 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j98rl" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.490396 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" event={"ID":"8a0b82ca-6250-4ca2-abb5-2704907e4f98","Type":"ContainerStarted","Data":"64a867aa78afd8daf9a1b616af7ef18dcf68afc5569b3cdb9961d3d6a6e905c8"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.490500 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" event={"ID":"8a0b82ca-6250-4ca2-abb5-2704907e4f98","Type":"ContainerStarted","Data":"886078628c1e4ef8135a4d92499ec7accaf44a7f06722e89414cf2e06c45726e"} Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.491420 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.497612 4990 scope.go:117] "RemoveContainer" containerID="02d693cb3abf43b38366de04886fea8d8fcf8aa59103b1ac5530b5cd332785e6" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.500619 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.540322 4990 scope.go:117] "RemoveContainer" containerID="daf286fe6684453d7969af2d97d03333d4c9d9106f671079d77dd8c342eae668" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.554182 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9n6hl" podStartSLOduration=2.554162834 podStartE2EDuration="2.554162834s" podCreationTimestamp="2025-12-03 12:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:44:51.522231069 +0000 UTC m=+439.664142308" watchObservedRunningTime="2025-12-03 12:44:51.554162834 +0000 UTC m=+439.696074063" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.579585 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jxn9m"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.583379 4990 scope.go:117] "RemoveContainer" containerID="f2eece2293c6748c775dc6db15f514023cd3c193ffc2180cec5c79f1f47b4539" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.595802 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jxn9m"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.603824 4990 scope.go:117] "RemoveContainer" containerID="abee2296d7ec88b389ee90728338efabe2412217d9fff8f741ac204450ba01f0" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.603985 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zn87g"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.616910 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zn87g"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.621058 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gmsz9"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.627978 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gmsz9"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.629907 4990 scope.go:117] "RemoveContainer" containerID="72aad500c91a6505c1c7d762fc1a45efaaa852347212d9c2cf391a8ae7a7b09f" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.632991 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbqqz"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.638601 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbqqz"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.644152 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j98rl"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.650557 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j98rl"] Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.653681 4990 scope.go:117] "RemoveContainer" containerID="3f98eaa052dcdb6fa554f07427a51b5a33183160e271e246fb2e3578b2f0a653" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.681648 4990 scope.go:117] "RemoveContainer" containerID="98ce943cfb69275a66c986eafff99bc65a3a80144327c59fb8fa0c49f9df9f23" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.699085 4990 scope.go:117] "RemoveContainer" containerID="2b9b4f697a3b4a26f7dd6745d64d07970e5e1e5040d084b650145ac110a347d8" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.716233 4990 scope.go:117] "RemoveContainer" containerID="ddb608ddbd922ae62062a0046c520b17134120f655d85c23fd62f4b391bb5145" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.736272 4990 scope.go:117] "RemoveContainer" containerID="8f42b33dc07af86a83a56129107ca063742b5b604584ab5cfb4584c16ef8fc3a" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.753860 4990 scope.go:117] "RemoveContainer" containerID="2a1485703ee2e7f0106e40143e7b7187d8344aa0fa2637a341d46dceccb44993" Dec 03 12:44:51 crc kubenswrapper[4990]: I1203 12:44:51.820490 4990 scope.go:117] "RemoveContainer" containerID="47e5ddb22e46cb429b080d91d979ec00a59f7a5d70be5f70030f50c5e950310e" Dec 03 12:44:52 crc kubenswrapper[4990]: I1203 12:44:52.276658 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" path="/var/lib/kubelet/pods/2870df3c-124c-4a2d-8f88-4fec5f65740c/volumes" Dec 03 12:44:52 crc kubenswrapper[4990]: I1203 12:44:52.277916 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" path="/var/lib/kubelet/pods/34bb2446-2a47-4b35-b5a1-13d8de74822e/volumes" Dec 03 12:44:52 crc kubenswrapper[4990]: I1203 12:44:52.278706 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8108de1d-75ea-419d-9658-e5a035088b36" path="/var/lib/kubelet/pods/8108de1d-75ea-419d-9658-e5a035088b36/volumes" Dec 03 12:44:52 crc kubenswrapper[4990]: I1203 12:44:52.280334 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" path="/var/lib/kubelet/pods/89cb569a-c36f-481c-b5a0-00694046c00c/volumes" Dec 03 12:44:52 crc kubenswrapper[4990]: I1203 12:44:52.281234 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" path="/var/lib/kubelet/pods/ddec4b1a-b34e-45f7-95e6-a04d29dde82c/volumes" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303344 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xbrs"] Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303593 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303610 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303629 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303636 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303645 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303652 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303660 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303666 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303674 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303680 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303689 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303694 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303703 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303710 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303720 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303725 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303733 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303739 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303748 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303754 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="extract-content" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303763 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303770 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303779 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303785 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303795 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303802 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.303811 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303817 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="extract-utilities" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303909 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddec4b1a-b34e-45f7-95e6-a04d29dde82c" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303916 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="34bb2446-2a47-4b35-b5a1-13d8de74822e" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303924 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="89cb569a-c36f-481c-b5a0-00694046c00c" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303932 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303942 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8108de1d-75ea-419d-9658-e5a035088b36" containerName="registry-server" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.303957 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: E1203 12:44:53.304060 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.304068 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.304143 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2870df3c-124c-4a2d-8f88-4fec5f65740c" containerName="marketplace-operator" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.304789 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.310092 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.365672 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xbrs"] Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.407429 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd9vk\" (UniqueName: \"kubernetes.io/projected/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-kube-api-access-cd9vk\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.408015 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-utilities\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.408096 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-catalog-content\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.505159 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zcrqv"] Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.506378 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.509057 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-catalog-content\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.509135 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd9vk\" (UniqueName: \"kubernetes.io/projected/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-kube-api-access-cd9vk\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.509221 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-utilities\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.509880 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-utilities\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.509973 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-catalog-content\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.510541 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.528256 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zcrqv"] Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.552425 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd9vk\" (UniqueName: \"kubernetes.io/projected/0b2e5cc6-9caa-458a-9f33-e0957c8fc435-kube-api-access-cd9vk\") pod \"certified-operators-6xbrs\" (UID: \"0b2e5cc6-9caa-458a-9f33-e0957c8fc435\") " pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.610535 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnlzm\" (UniqueName: \"kubernetes.io/projected/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-kube-api-access-dnlzm\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.610633 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-utilities\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.610832 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-catalog-content\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.624499 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.711906 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-utilities\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.711961 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-catalog-content\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.712034 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnlzm\" (UniqueName: \"kubernetes.io/projected/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-kube-api-access-dnlzm\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.712828 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-utilities\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.713079 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-catalog-content\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.747708 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnlzm\" (UniqueName: \"kubernetes.io/projected/88e0104c-98da-4fd8-86c6-5ba7a904d2b5-kube-api-access-dnlzm\") pod \"community-operators-zcrqv\" (UID: \"88e0104c-98da-4fd8-86c6-5ba7a904d2b5\") " pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:53 crc kubenswrapper[4990]: I1203 12:44:53.832560 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.078716 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xbrs"] Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.287377 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zcrqv"] Dec 03 12:44:54 crc kubenswrapper[4990]: W1203 12:44:54.295806 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88e0104c_98da_4fd8_86c6_5ba7a904d2b5.slice/crio-9b2acab2f84e3f53743975936ff2af758c4eeef548c97c449233a17573833abc WatchSource:0}: Error finding container 9b2acab2f84e3f53743975936ff2af758c4eeef548c97c449233a17573833abc: Status 404 returned error can't find the container with id 9b2acab2f84e3f53743975936ff2af758c4eeef548c97c449233a17573833abc Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.528006 4990 generic.go:334] "Generic (PLEG): container finished" podID="88e0104c-98da-4fd8-86c6-5ba7a904d2b5" containerID="345ef394ae7cf169dcd4389c6a503e9f022d6d4998d5596a4df1469a19d90c44" exitCode=0 Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.528122 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcrqv" event={"ID":"88e0104c-98da-4fd8-86c6-5ba7a904d2b5","Type":"ContainerDied","Data":"345ef394ae7cf169dcd4389c6a503e9f022d6d4998d5596a4df1469a19d90c44"} Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.528161 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcrqv" event={"ID":"88e0104c-98da-4fd8-86c6-5ba7a904d2b5","Type":"ContainerStarted","Data":"9b2acab2f84e3f53743975936ff2af758c4eeef548c97c449233a17573833abc"} Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.535385 4990 generic.go:334] "Generic (PLEG): container finished" podID="0b2e5cc6-9caa-458a-9f33-e0957c8fc435" containerID="454e80699d79c40db09f25449d1053eeec7b46865dd9a51096fe5c605cea381d" exitCode=0 Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.535469 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbrs" event={"ID":"0b2e5cc6-9caa-458a-9f33-e0957c8fc435","Type":"ContainerDied","Data":"454e80699d79c40db09f25449d1053eeec7b46865dd9a51096fe5c605cea381d"} Dec 03 12:44:54 crc kubenswrapper[4990]: I1203 12:44:54.537309 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbrs" event={"ID":"0b2e5cc6-9caa-458a-9f33-e0957c8fc435","Type":"ContainerStarted","Data":"a56415bf87e359d0933d0c999e434a5e1fb44c185cc6bc53990755a6b6d08ec3"} Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.706575 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pcfbg"] Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.709809 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.716377 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.727435 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcfbg"] Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.846541 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-utilities\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.846679 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-catalog-content\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.846772 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqwf9\" (UniqueName: \"kubernetes.io/projected/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-kube-api-access-sqwf9\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.904402 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h4nph"] Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.905768 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.910680 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.914961 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h4nph"] Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.947820 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-catalog-content\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.948128 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crlnp\" (UniqueName: \"kubernetes.io/projected/32457d7b-2669-40b2-8b4c-6d15db06c596-kube-api-access-crlnp\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.948231 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqwf9\" (UniqueName: \"kubernetes.io/projected/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-kube-api-access-sqwf9\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.948366 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32457d7b-2669-40b2-8b4c-6d15db06c596-catalog-content\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.948473 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-utilities\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.948545 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-catalog-content\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.948577 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32457d7b-2669-40b2-8b4c-6d15db06c596-utilities\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.949172 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-utilities\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:55 crc kubenswrapper[4990]: I1203 12:44:55.973900 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqwf9\" (UniqueName: \"kubernetes.io/projected/2b52b9cb-d5fd-4916-a51b-440f95d47bbd-kube-api-access-sqwf9\") pod \"redhat-marketplace-pcfbg\" (UID: \"2b52b9cb-d5fd-4916-a51b-440f95d47bbd\") " pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.050823 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crlnp\" (UniqueName: \"kubernetes.io/projected/32457d7b-2669-40b2-8b4c-6d15db06c596-kube-api-access-crlnp\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.050902 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32457d7b-2669-40b2-8b4c-6d15db06c596-catalog-content\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.050928 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32457d7b-2669-40b2-8b4c-6d15db06c596-utilities\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.051596 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32457d7b-2669-40b2-8b4c-6d15db06c596-utilities\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.051731 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32457d7b-2669-40b2-8b4c-6d15db06c596-catalog-content\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.083446 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crlnp\" (UniqueName: \"kubernetes.io/projected/32457d7b-2669-40b2-8b4c-6d15db06c596-kube-api-access-crlnp\") pod \"redhat-operators-h4nph\" (UID: \"32457d7b-2669-40b2-8b4c-6d15db06c596\") " pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.119983 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.234728 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.554691 4990 generic.go:334] "Generic (PLEG): container finished" podID="0b2e5cc6-9caa-458a-9f33-e0957c8fc435" containerID="d624b12fcffa645ebb7a24fbbc78cc6ba489b1c14d5193f8111f1614de54dcd4" exitCode=0 Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.554787 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbrs" event={"ID":"0b2e5cc6-9caa-458a-9f33-e0957c8fc435","Type":"ContainerDied","Data":"d624b12fcffa645ebb7a24fbbc78cc6ba489b1c14d5193f8111f1614de54dcd4"} Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.559105 4990 generic.go:334] "Generic (PLEG): container finished" podID="88e0104c-98da-4fd8-86c6-5ba7a904d2b5" containerID="6cd4a7c4e363ee22f33fc5e91b6251ff0b25083769b3ab570c79472baf7ba8c5" exitCode=0 Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.559158 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcrqv" event={"ID":"88e0104c-98da-4fd8-86c6-5ba7a904d2b5","Type":"ContainerDied","Data":"6cd4a7c4e363ee22f33fc5e91b6251ff0b25083769b3ab570c79472baf7ba8c5"} Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.662264 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcfbg"] Dec 03 12:44:56 crc kubenswrapper[4990]: W1203 12:44:56.672186 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b52b9cb_d5fd_4916_a51b_440f95d47bbd.slice/crio-9c2c7253a1ece0e146b26078de32ece8fd08d37c7113606c07964fac4de03fe1 WatchSource:0}: Error finding container 9c2c7253a1ece0e146b26078de32ece8fd08d37c7113606c07964fac4de03fe1: Status 404 returned error can't find the container with id 9c2c7253a1ece0e146b26078de32ece8fd08d37c7113606c07964fac4de03fe1 Dec 03 12:44:56 crc kubenswrapper[4990]: I1203 12:44:56.781805 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h4nph"] Dec 03 12:44:57 crc kubenswrapper[4990]: I1203 12:44:57.567694 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4nph" event={"ID":"32457d7b-2669-40b2-8b4c-6d15db06c596","Type":"ContainerStarted","Data":"785c8c50a92621efb66da801339ebf6fd2967c46e540780bb904906b2e774e43"} Dec 03 12:44:57 crc kubenswrapper[4990]: I1203 12:44:57.570122 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcfbg" event={"ID":"2b52b9cb-d5fd-4916-a51b-440f95d47bbd","Type":"ContainerStarted","Data":"9c2c7253a1ece0e146b26078de32ece8fd08d37c7113606c07964fac4de03fe1"} Dec 03 12:44:58 crc kubenswrapper[4990]: I1203 12:44:58.577659 4990 generic.go:334] "Generic (PLEG): container finished" podID="2b52b9cb-d5fd-4916-a51b-440f95d47bbd" containerID="1d7fb7f4f529c8b1d9814b294ba3953fd4026d85055dde834304bdbb2014210d" exitCode=0 Dec 03 12:44:58 crc kubenswrapper[4990]: I1203 12:44:58.577892 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcfbg" event={"ID":"2b52b9cb-d5fd-4916-a51b-440f95d47bbd","Type":"ContainerDied","Data":"1d7fb7f4f529c8b1d9814b294ba3953fd4026d85055dde834304bdbb2014210d"} Dec 03 12:44:58 crc kubenswrapper[4990]: I1203 12:44:58.579610 4990 generic.go:334] "Generic (PLEG): container finished" podID="32457d7b-2669-40b2-8b4c-6d15db06c596" containerID="1646451c4012b7459543bfa70fa63b05b14e1396278d1d9a249205f36d9ff9ea" exitCode=0 Dec 03 12:44:58 crc kubenswrapper[4990]: I1203 12:44:58.579638 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4nph" event={"ID":"32457d7b-2669-40b2-8b4c-6d15db06c596","Type":"ContainerDied","Data":"1646451c4012b7459543bfa70fa63b05b14e1396278d1d9a249205f36d9ff9ea"} Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.196657 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h"] Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.203815 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.207003 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h"] Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.207291 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.207647 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.329566 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/471bdda0-0146-48d8-af64-0cfd178fdd48-config-volume\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.330194 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/471bdda0-0146-48d8-af64-0cfd178fdd48-secret-volume\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.330259 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85zsn\" (UniqueName: \"kubernetes.io/projected/471bdda0-0146-48d8-af64-0cfd178fdd48-kube-api-access-85zsn\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.431928 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85zsn\" (UniqueName: \"kubernetes.io/projected/471bdda0-0146-48d8-af64-0cfd178fdd48-kube-api-access-85zsn\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.432034 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/471bdda0-0146-48d8-af64-0cfd178fdd48-config-volume\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.432069 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/471bdda0-0146-48d8-af64-0cfd178fdd48-secret-volume\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.433831 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/471bdda0-0146-48d8-af64-0cfd178fdd48-config-volume\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.439623 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/471bdda0-0146-48d8-af64-0cfd178fdd48-secret-volume\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:00 crc kubenswrapper[4990]: I1203 12:45:00.456715 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85zsn\" (UniqueName: \"kubernetes.io/projected/471bdda0-0146-48d8-af64-0cfd178fdd48-kube-api-access-85zsn\") pod \"collect-profiles-29412765-jzt4h\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:01 crc kubenswrapper[4990]: I1203 12:45:01.365104 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:01 crc kubenswrapper[4990]: I1203 12:45:01.820697 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h"] Dec 03 12:45:02 crc kubenswrapper[4990]: I1203 12:45:02.625894 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" event={"ID":"471bdda0-0146-48d8-af64-0cfd178fdd48","Type":"ContainerStarted","Data":"51b1a7437c7ae8b4a4f8972147db3c55a1c226b1ffb24e9926fb138821602ebd"} Dec 03 12:45:03 crc kubenswrapper[4990]: I1203 12:45:03.633778 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zcrqv" event={"ID":"88e0104c-98da-4fd8-86c6-5ba7a904d2b5","Type":"ContainerStarted","Data":"13009f0d2ea9a1971ec34e950382bf5b41cf48d6d6200983599d5ac3905adfe5"} Dec 03 12:45:04 crc kubenswrapper[4990]: I1203 12:45:04.645614 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" event={"ID":"471bdda0-0146-48d8-af64-0cfd178fdd48","Type":"ContainerStarted","Data":"7f48c4ef412c45a8d91880a69311c45ef0381639de6bad984188b4f929a418ba"} Dec 03 12:45:04 crc kubenswrapper[4990]: I1203 12:45:04.651484 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbrs" event={"ID":"0b2e5cc6-9caa-458a-9f33-e0957c8fc435","Type":"ContainerStarted","Data":"df7b15ed00cee60b2856ce426ae053d8d88a9051bc22f7bf6efb9d7bedaee55d"} Dec 03 12:45:04 crc kubenswrapper[4990]: I1203 12:45:04.672303 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" podStartSLOduration=4.672283879 podStartE2EDuration="4.672283879s" podCreationTimestamp="2025-12-03 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:45:04.666777713 +0000 UTC m=+452.808688952" watchObservedRunningTime="2025-12-03 12:45:04.672283879 +0000 UTC m=+452.814195108" Dec 03 12:45:04 crc kubenswrapper[4990]: I1203 12:45:04.694712 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zcrqv" podStartSLOduration=6.253868224 podStartE2EDuration="11.694681542s" podCreationTimestamp="2025-12-03 12:44:53 +0000 UTC" firstStartedPulling="2025-12-03 12:44:54.530086436 +0000 UTC m=+442.671997665" lastFinishedPulling="2025-12-03 12:44:59.970899754 +0000 UTC m=+448.112810983" observedRunningTime="2025-12-03 12:45:04.691890918 +0000 UTC m=+452.833802167" watchObservedRunningTime="2025-12-03 12:45:04.694681542 +0000 UTC m=+452.836592761" Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.659676 4990 generic.go:334] "Generic (PLEG): container finished" podID="471bdda0-0146-48d8-af64-0cfd178fdd48" containerID="7f48c4ef412c45a8d91880a69311c45ef0381639de6bad984188b4f929a418ba" exitCode=0 Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.660017 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" event={"ID":"471bdda0-0146-48d8-af64-0cfd178fdd48","Type":"ContainerDied","Data":"7f48c4ef412c45a8d91880a69311c45ef0381639de6bad984188b4f929a418ba"} Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.662092 4990 generic.go:334] "Generic (PLEG): container finished" podID="2b52b9cb-d5fd-4916-a51b-440f95d47bbd" containerID="2f6cfe6c19e7ed92e6c6213a58416fbbae2021cd5f78864ecd768e28f907e3b5" exitCode=0 Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.662185 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcfbg" event={"ID":"2b52b9cb-d5fd-4916-a51b-440f95d47bbd","Type":"ContainerDied","Data":"2f6cfe6c19e7ed92e6c6213a58416fbbae2021cd5f78864ecd768e28f907e3b5"} Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.665633 4990 generic.go:334] "Generic (PLEG): container finished" podID="32457d7b-2669-40b2-8b4c-6d15db06c596" containerID="090965ec5da5076423410f2280bbd81dbb63cdf75000491547a070e370e1278b" exitCode=0 Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.665716 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4nph" event={"ID":"32457d7b-2669-40b2-8b4c-6d15db06c596","Type":"ContainerDied","Data":"090965ec5da5076423410f2280bbd81dbb63cdf75000491547a070e370e1278b"} Dec 03 12:45:05 crc kubenswrapper[4990]: I1203 12:45:05.687429 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xbrs" podStartSLOduration=7.310531334 podStartE2EDuration="12.687403659s" podCreationTimestamp="2025-12-03 12:44:53 +0000 UTC" firstStartedPulling="2025-12-03 12:44:54.537291978 +0000 UTC m=+442.679203207" lastFinishedPulling="2025-12-03 12:44:59.914164313 +0000 UTC m=+448.056075532" observedRunningTime="2025-12-03 12:45:04.716024437 +0000 UTC m=+452.857935666" watchObservedRunningTime="2025-12-03 12:45:05.687403659 +0000 UTC m=+453.829314888" Dec 03 12:45:06 crc kubenswrapper[4990]: I1203 12:45:06.674125 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcfbg" event={"ID":"2b52b9cb-d5fd-4916-a51b-440f95d47bbd","Type":"ContainerStarted","Data":"d2dec3261aac76f207aa1a90cb6f8e39596ae1f719abebe3c3400a0879d5536f"} Dec 03 12:45:06 crc kubenswrapper[4990]: I1203 12:45:06.676882 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4nph" event={"ID":"32457d7b-2669-40b2-8b4c-6d15db06c596","Type":"ContainerStarted","Data":"1eeb0448adee21356a694390f3e89f3a3db2ec3627584aabf77e699e27ccdb70"} Dec 03 12:45:06 crc kubenswrapper[4990]: I1203 12:45:06.704727 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pcfbg" podStartSLOduration=6.209242673 podStartE2EDuration="11.704703257s" podCreationTimestamp="2025-12-03 12:44:55 +0000 UTC" firstStartedPulling="2025-12-03 12:45:00.606438957 +0000 UTC m=+448.748350186" lastFinishedPulling="2025-12-03 12:45:06.101899541 +0000 UTC m=+454.243810770" observedRunningTime="2025-12-03 12:45:06.702379066 +0000 UTC m=+454.844290295" watchObservedRunningTime="2025-12-03 12:45:06.704703257 +0000 UTC m=+454.846614486" Dec 03 12:45:06 crc kubenswrapper[4990]: I1203 12:45:06.723473 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h4nph" podStartSLOduration=6.152960673 podStartE2EDuration="11.723435563s" podCreationTimestamp="2025-12-03 12:44:55 +0000 UTC" firstStartedPulling="2025-12-03 12:45:00.606246322 +0000 UTC m=+448.748157551" lastFinishedPulling="2025-12-03 12:45:06.176721212 +0000 UTC m=+454.318632441" observedRunningTime="2025-12-03 12:45:06.720047534 +0000 UTC m=+454.861958783" watchObservedRunningTime="2025-12-03 12:45:06.723435563 +0000 UTC m=+454.865346792" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.101437 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.188052 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/471bdda0-0146-48d8-af64-0cfd178fdd48-config-volume\") pod \"471bdda0-0146-48d8-af64-0cfd178fdd48\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.188144 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85zsn\" (UniqueName: \"kubernetes.io/projected/471bdda0-0146-48d8-af64-0cfd178fdd48-kube-api-access-85zsn\") pod \"471bdda0-0146-48d8-af64-0cfd178fdd48\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.188252 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/471bdda0-0146-48d8-af64-0cfd178fdd48-secret-volume\") pod \"471bdda0-0146-48d8-af64-0cfd178fdd48\" (UID: \"471bdda0-0146-48d8-af64-0cfd178fdd48\") " Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.188984 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/471bdda0-0146-48d8-af64-0cfd178fdd48-config-volume" (OuterVolumeSpecName: "config-volume") pod "471bdda0-0146-48d8-af64-0cfd178fdd48" (UID: "471bdda0-0146-48d8-af64-0cfd178fdd48"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.246032 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/471bdda0-0146-48d8-af64-0cfd178fdd48-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "471bdda0-0146-48d8-af64-0cfd178fdd48" (UID: "471bdda0-0146-48d8-af64-0cfd178fdd48"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.246746 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/471bdda0-0146-48d8-af64-0cfd178fdd48-kube-api-access-85zsn" (OuterVolumeSpecName: "kube-api-access-85zsn") pod "471bdda0-0146-48d8-af64-0cfd178fdd48" (UID: "471bdda0-0146-48d8-af64-0cfd178fdd48"). InnerVolumeSpecName "kube-api-access-85zsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.290426 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/471bdda0-0146-48d8-af64-0cfd178fdd48-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.290500 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/471bdda0-0146-48d8-af64-0cfd178fdd48-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.290513 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85zsn\" (UniqueName: \"kubernetes.io/projected/471bdda0-0146-48d8-af64-0cfd178fdd48-kube-api-access-85zsn\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.684058 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" event={"ID":"471bdda0-0146-48d8-af64-0cfd178fdd48","Type":"ContainerDied","Data":"51b1a7437c7ae8b4a4f8972147db3c55a1c226b1ffb24e9926fb138821602ebd"} Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.684107 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51b1a7437c7ae8b4a4f8972147db3c55a1c226b1ffb24e9926fb138821602ebd" Dec 03 12:45:07 crc kubenswrapper[4990]: I1203 12:45:07.684168 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h" Dec 03 12:45:12 crc kubenswrapper[4990]: I1203 12:45:12.888164 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" podUID="791d5adb-b1b5-4cf6-a003-f8ad41da334e" containerName="registry" containerID="cri-o://50a28a3f7d4be32638a809532df45f870241910d4e0897677fe9664cd5100c0d" gracePeriod=30 Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.626155 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.626222 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.670361 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.763879 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xbrs" Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.833239 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.833395 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:45:13 crc kubenswrapper[4990]: I1203 12:45:13.875610 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:45:14 crc kubenswrapper[4990]: I1203 12:45:14.771316 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zcrqv" Dec 03 12:45:15 crc kubenswrapper[4990]: I1203 12:45:15.733618 4990 generic.go:334] "Generic (PLEG): container finished" podID="791d5adb-b1b5-4cf6-a003-f8ad41da334e" containerID="50a28a3f7d4be32638a809532df45f870241910d4e0897677fe9664cd5100c0d" exitCode=0 Dec 03 12:45:15 crc kubenswrapper[4990]: I1203 12:45:15.733702 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" event={"ID":"791d5adb-b1b5-4cf6-a003-f8ad41da334e","Type":"ContainerDied","Data":"50a28a3f7d4be32638a809532df45f870241910d4e0897677fe9664cd5100c0d"} Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.104820 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.121310 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.121378 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.175396 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.235907 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.236007 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240064 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-trusted-ca\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240190 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-tls\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240220 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/791d5adb-b1b5-4cf6-a003-f8ad41da334e-installation-pull-secrets\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240248 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-bound-sa-token\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240279 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-certificates\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240307 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqzxd\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-kube-api-access-fqzxd\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240501 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.240570 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/791d5adb-b1b5-4cf6-a003-f8ad41da334e-ca-trust-extracted\") pod \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\" (UID: \"791d5adb-b1b5-4cf6-a003-f8ad41da334e\") " Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.243089 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.244121 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.251176 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.251348 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-kube-api-access-fqzxd" (OuterVolumeSpecName: "kube-api-access-fqzxd") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "kube-api-access-fqzxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.251634 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791d5adb-b1b5-4cf6-a003-f8ad41da334e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.256891 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.258138 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.263328 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/791d5adb-b1b5-4cf6-a003-f8ad41da334e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "791d5adb-b1b5-4cf6-a003-f8ad41da334e" (UID: "791d5adb-b1b5-4cf6-a003-f8ad41da334e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.278751 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342858 4990 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342903 4990 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/791d5adb-b1b5-4cf6-a003-f8ad41da334e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342925 4990 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342938 4990 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342950 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqzxd\" (UniqueName: \"kubernetes.io/projected/791d5adb-b1b5-4cf6-a003-f8ad41da334e-kube-api-access-fqzxd\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342962 4990 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/791d5adb-b1b5-4cf6-a003-f8ad41da334e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.342973 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/791d5adb-b1b5-4cf6-a003-f8ad41da334e-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.741111 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" event={"ID":"791d5adb-b1b5-4cf6-a003-f8ad41da334e","Type":"ContainerDied","Data":"ea9073754fadc6a03a7bbb142eb2d380cfaca19c9f945daa6d63f44e4dec6b58"} Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.741195 4990 scope.go:117] "RemoveContainer" containerID="50a28a3f7d4be32638a809532df45f870241910d4e0897677fe9664cd5100c0d" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.741316 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-twmj9" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.767172 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-twmj9"] Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.773248 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-twmj9"] Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.786892 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h4nph" Dec 03 12:45:16 crc kubenswrapper[4990]: I1203 12:45:16.787388 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pcfbg" Dec 03 12:45:18 crc kubenswrapper[4990]: I1203 12:45:18.272608 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="791d5adb-b1b5-4cf6-a003-f8ad41da334e" path="/var/lib/kubelet/pods/791d5adb-b1b5-4cf6-a003-f8ad41da334e/volumes" Dec 03 12:46:33 crc kubenswrapper[4990]: I1203 12:46:33.287435 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:46:33 crc kubenswrapper[4990]: I1203 12:46:33.288116 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:47:03 crc kubenswrapper[4990]: I1203 12:47:03.287151 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:47:03 crc kubenswrapper[4990]: I1203 12:47:03.288274 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.286331 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.288158 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.288265 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.288981 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"daf99345f123e10ced8b9ee63d4b06d27556fb355d285a58334782f921bd8551"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.289117 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://daf99345f123e10ced8b9ee63d4b06d27556fb355d285a58334782f921bd8551" gracePeriod=600 Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.655434 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="daf99345f123e10ced8b9ee63d4b06d27556fb355d285a58334782f921bd8551" exitCode=0 Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.655524 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"daf99345f123e10ced8b9ee63d4b06d27556fb355d285a58334782f921bd8551"} Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.655567 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"0e6f8ca4fb394aa93d3f85ed97825d9ed55398588ce9a009c2c47f07ca7524e8"} Dec 03 12:47:33 crc kubenswrapper[4990]: I1203 12:47:33.655589 4990 scope.go:117] "RemoveContainer" containerID="8ff22e0abfe7a9119c84cea1563bea69a1449b575844b4f78fdcfa08778a9149" Dec 03 12:49:33 crc kubenswrapper[4990]: I1203 12:49:33.287033 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:49:33 crc kubenswrapper[4990]: I1203 12:49:33.287959 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:50:03 crc kubenswrapper[4990]: I1203 12:50:03.286907 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:50:03 crc kubenswrapper[4990]: I1203 12:50:03.287850 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:50:26 crc kubenswrapper[4990]: I1203 12:50:26.618856 4990 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 12:50:33 crc kubenswrapper[4990]: I1203 12:50:33.286699 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:50:33 crc kubenswrapper[4990]: I1203 12:50:33.287746 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:50:33 crc kubenswrapper[4990]: I1203 12:50:33.287835 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:50:33 crc kubenswrapper[4990]: I1203 12:50:33.288776 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0e6f8ca4fb394aa93d3f85ed97825d9ed55398588ce9a009c2c47f07ca7524e8"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 12:50:33 crc kubenswrapper[4990]: I1203 12:50:33.288865 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://0e6f8ca4fb394aa93d3f85ed97825d9ed55398588ce9a009c2c47f07ca7524e8" gracePeriod=600 Dec 03 12:50:34 crc kubenswrapper[4990]: I1203 12:50:34.275694 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="0e6f8ca4fb394aa93d3f85ed97825d9ed55398588ce9a009c2c47f07ca7524e8" exitCode=0 Dec 03 12:50:34 crc kubenswrapper[4990]: I1203 12:50:34.275775 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"0e6f8ca4fb394aa93d3f85ed97825d9ed55398588ce9a009c2c47f07ca7524e8"} Dec 03 12:50:34 crc kubenswrapper[4990]: I1203 12:50:34.276225 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"ca4307683ea7333a625e2203775fa26ab0df3b188bf9bac1cd9b6194d3df672f"} Dec 03 12:50:34 crc kubenswrapper[4990]: I1203 12:50:34.276261 4990 scope.go:117] "RemoveContainer" containerID="daf99345f123e10ced8b9ee63d4b06d27556fb355d285a58334782f921bd8551" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.167041 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xrqmv"] Dec 03 12:51:23 crc kubenswrapper[4990]: E1203 12:51:23.167924 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471bdda0-0146-48d8-af64-0cfd178fdd48" containerName="collect-profiles" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.167942 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="471bdda0-0146-48d8-af64-0cfd178fdd48" containerName="collect-profiles" Dec 03 12:51:23 crc kubenswrapper[4990]: E1203 12:51:23.167966 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791d5adb-b1b5-4cf6-a003-f8ad41da334e" containerName="registry" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.167976 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="791d5adb-b1b5-4cf6-a003-f8ad41da334e" containerName="registry" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.168087 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="471bdda0-0146-48d8-af64-0cfd178fdd48" containerName="collect-profiles" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.168109 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="791d5adb-b1b5-4cf6-a003-f8ad41da334e" containerName="registry" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.168658 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xrqmv" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.170534 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.170864 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.171672 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xlfdw"] Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.172544 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.174853 4990 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vkswb" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.177867 4990 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qld9t" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.192269 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xlfdw"] Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.199232 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xrqmv"] Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.203579 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8ghd6"] Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.204440 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.207814 4990 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-dnslg" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.211592 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8ghd6"] Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.365578 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmrt5\" (UniqueName: \"kubernetes.io/projected/9dd6b4f0-81be-480c-84d6-22ba5b462ab0-kube-api-access-qmrt5\") pod \"cert-manager-cainjector-7f985d654d-xlfdw\" (UID: \"9dd6b4f0-81be-480c-84d6-22ba5b462ab0\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.365682 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5rpw\" (UniqueName: \"kubernetes.io/projected/555f5cb7-9649-4c3a-8b77-244edcb83c7f-kube-api-access-c5rpw\") pod \"cert-manager-5b446d88c5-xrqmv\" (UID: \"555f5cb7-9649-4c3a-8b77-244edcb83c7f\") " pod="cert-manager/cert-manager-5b446d88c5-xrqmv" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.365726 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7wbg\" (UniqueName: \"kubernetes.io/projected/3196eb4c-4adf-493f-ade6-a9c839d0e4e6-kube-api-access-m7wbg\") pod \"cert-manager-webhook-5655c58dd6-8ghd6\" (UID: \"3196eb4c-4adf-493f-ade6-a9c839d0e4e6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.468322 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmrt5\" (UniqueName: \"kubernetes.io/projected/9dd6b4f0-81be-480c-84d6-22ba5b462ab0-kube-api-access-qmrt5\") pod \"cert-manager-cainjector-7f985d654d-xlfdw\" (UID: \"9dd6b4f0-81be-480c-84d6-22ba5b462ab0\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.468750 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5rpw\" (UniqueName: \"kubernetes.io/projected/555f5cb7-9649-4c3a-8b77-244edcb83c7f-kube-api-access-c5rpw\") pod \"cert-manager-5b446d88c5-xrqmv\" (UID: \"555f5cb7-9649-4c3a-8b77-244edcb83c7f\") " pod="cert-manager/cert-manager-5b446d88c5-xrqmv" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.468902 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7wbg\" (UniqueName: \"kubernetes.io/projected/3196eb4c-4adf-493f-ade6-a9c839d0e4e6-kube-api-access-m7wbg\") pod \"cert-manager-webhook-5655c58dd6-8ghd6\" (UID: \"3196eb4c-4adf-493f-ade6-a9c839d0e4e6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.503677 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7wbg\" (UniqueName: \"kubernetes.io/projected/3196eb4c-4adf-493f-ade6-a9c839d0e4e6-kube-api-access-m7wbg\") pod \"cert-manager-webhook-5655c58dd6-8ghd6\" (UID: \"3196eb4c-4adf-493f-ade6-a9c839d0e4e6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.503720 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmrt5\" (UniqueName: \"kubernetes.io/projected/9dd6b4f0-81be-480c-84d6-22ba5b462ab0-kube-api-access-qmrt5\") pod \"cert-manager-cainjector-7f985d654d-xlfdw\" (UID: \"9dd6b4f0-81be-480c-84d6-22ba5b462ab0\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.506563 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5rpw\" (UniqueName: \"kubernetes.io/projected/555f5cb7-9649-4c3a-8b77-244edcb83c7f-kube-api-access-c5rpw\") pod \"cert-manager-5b446d88c5-xrqmv\" (UID: \"555f5cb7-9649-4c3a-8b77-244edcb83c7f\") " pod="cert-manager/cert-manager-5b446d88c5-xrqmv" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.531322 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.544358 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.788370 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xrqmv" Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.821126 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xlfdw"] Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.846166 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 12:51:23 crc kubenswrapper[4990]: I1203 12:51:23.868318 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8ghd6"] Dec 03 12:51:24 crc kubenswrapper[4990]: I1203 12:51:24.036312 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xrqmv"] Dec 03 12:51:24 crc kubenswrapper[4990]: W1203 12:51:24.043084 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod555f5cb7_9649_4c3a_8b77_244edcb83c7f.slice/crio-d4f3a5e614a15031ad76fa8aab21ac97d611c6464b729b26584f11492f138f43 WatchSource:0}: Error finding container d4f3a5e614a15031ad76fa8aab21ac97d611c6464b729b26584f11492f138f43: Status 404 returned error can't find the container with id d4f3a5e614a15031ad76fa8aab21ac97d611c6464b729b26584f11492f138f43 Dec 03 12:51:24 crc kubenswrapper[4990]: I1203 12:51:24.412000 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" event={"ID":"9dd6b4f0-81be-480c-84d6-22ba5b462ab0","Type":"ContainerStarted","Data":"1b9981055e6992407bc4a99d66878b421624488675dd0df22d416a7f7fd41eb6"} Dec 03 12:51:24 crc kubenswrapper[4990]: I1203 12:51:24.413681 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xrqmv" event={"ID":"555f5cb7-9649-4c3a-8b77-244edcb83c7f","Type":"ContainerStarted","Data":"d4f3a5e614a15031ad76fa8aab21ac97d611c6464b729b26584f11492f138f43"} Dec 03 12:51:24 crc kubenswrapper[4990]: I1203 12:51:24.414622 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" event={"ID":"3196eb4c-4adf-493f-ade6-a9c839d0e4e6","Type":"ContainerStarted","Data":"0ea379c0e4897d52e764582528c95b57d713dcba7bd0ee49474c78458f39b023"} Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.454494 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xrqmv" event={"ID":"555f5cb7-9649-4c3a-8b77-244edcb83c7f","Type":"ContainerStarted","Data":"d179a5329a124afc6e4e16d850e8e7837aaf6386bf90e607be55edec849d2cd9"} Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.457964 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" event={"ID":"3196eb4c-4adf-493f-ade6-a9c839d0e4e6","Type":"ContainerStarted","Data":"c812beeeab29f707169ad7742c1d7f71e09550c0e4ae5601a44b9f93b5e3c657"} Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.458064 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.459519 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" event={"ID":"9dd6b4f0-81be-480c-84d6-22ba5b462ab0","Type":"ContainerStarted","Data":"95ca987b9395fdc11f8a5d778f2ce84022ac998dc2d5398816c7020fab958b24"} Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.477223 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-xrqmv" podStartSLOduration=1.8666275959999998 podStartE2EDuration="6.477206331s" podCreationTimestamp="2025-12-03 12:51:23 +0000 UTC" firstStartedPulling="2025-12-03 12:51:24.046064855 +0000 UTC m=+832.187976084" lastFinishedPulling="2025-12-03 12:51:28.65664359 +0000 UTC m=+836.798554819" observedRunningTime="2025-12-03 12:51:29.472641 +0000 UTC m=+837.614552229" watchObservedRunningTime="2025-12-03 12:51:29.477206331 +0000 UTC m=+837.619117560" Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.497222 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" podStartSLOduration=1.8360765799999998 podStartE2EDuration="6.497204567s" podCreationTimestamp="2025-12-03 12:51:23 +0000 UTC" firstStartedPulling="2025-12-03 12:51:23.87731529 +0000 UTC m=+832.019226519" lastFinishedPulling="2025-12-03 12:51:28.538443277 +0000 UTC m=+836.680354506" observedRunningTime="2025-12-03 12:51:29.494034564 +0000 UTC m=+837.635945793" watchObservedRunningTime="2025-12-03 12:51:29.497204567 +0000 UTC m=+837.639115796" Dec 03 12:51:29 crc kubenswrapper[4990]: I1203 12:51:29.518545 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-xlfdw" podStartSLOduration=1.826465218 podStartE2EDuration="6.518527289s" podCreationTimestamp="2025-12-03 12:51:23 +0000 UTC" firstStartedPulling="2025-12-03 12:51:23.84576985 +0000 UTC m=+831.987681079" lastFinishedPulling="2025-12-03 12:51:28.537831921 +0000 UTC m=+836.679743150" observedRunningTime="2025-12-03 12:51:29.51290034 +0000 UTC m=+837.654811569" watchObservedRunningTime="2025-12-03 12:51:29.518527289 +0000 UTC m=+837.660438518" Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.877833 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7lrq6"] Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.879023 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-controller" containerID="cri-o://122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.879419 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.879405 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="northd" containerID="cri-o://387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.879525 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-acl-logging" containerID="cri-o://cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.879510 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-node" containerID="cri-o://c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.879834 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="sbdb" containerID="cri-o://10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.880030 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="nbdb" containerID="cri-o://00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: I1203 12:51:32.916571 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" containerID="cri-o://06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902" gracePeriod=30 Dec 03 12:51:32 crc kubenswrapper[4990]: E1203 12:51:32.982954 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5d4bd47_2fa6_4e6a_9697_9af4e227afe1.slice/crio-c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5d4bd47_2fa6_4e6a_9697_9af4e227afe1.slice/crio-122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe.scope\": RecentStats: unable to find data in memory cache]" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.486116 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/2.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.487139 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/1.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.487267 4990 generic.go:334] "Generic (PLEG): container finished" podID="41a42811-350f-4311-bb49-f7f116aa6952" containerID="82d1c99f98ceb0301fbb2c407c935e3b0fd5c83937c77e454284191b586dc10c" exitCode=2 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.487387 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerDied","Data":"82d1c99f98ceb0301fbb2c407c935e3b0fd5c83937c77e454284191b586dc10c"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.488028 4990 scope.go:117] "RemoveContainer" containerID="3701db16d65edb308cac94fa8425f7f81f55f71354fdc4d1c9893d6a8c7805ea" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.488567 4990 scope.go:117] "RemoveContainer" containerID="82d1c99f98ceb0301fbb2c407c935e3b0fd5c83937c77e454284191b586dc10c" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.492032 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovnkube-controller/3.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.495715 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovn-acl-logging/0.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496315 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovn-controller/0.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496780 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902" exitCode=0 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496812 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729" exitCode=0 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496822 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b" exitCode=0 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496833 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d" exitCode=0 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496844 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258" exitCode=143 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496854 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe" exitCode=143 Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496865 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496925 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496945 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496959 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496977 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.496992 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe"} Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.538781 4990 scope.go:117] "RemoveContainer" containerID="5744655b630fddf9e5bea7308d9eff4da0ca118579ef9f849536e0f5a4851fe9" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.548015 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-8ghd6" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.594391 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovn-acl-logging/0.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.595225 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovn-controller/0.log" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.595779 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.657594 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hnlgs"] Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658244 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658261 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658273 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="northd" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658283 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="northd" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658295 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="sbdb" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658303 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="sbdb" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658310 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-acl-logging" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658317 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-acl-logging" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658336 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kubecfg-setup" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658344 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kubecfg-setup" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658354 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="nbdb" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658361 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="nbdb" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658371 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-node" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658377 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-node" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658386 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658393 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658402 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658408 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658415 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658421 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658431 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658438 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658563 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658572 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658580 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="sbdb" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658591 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-acl-logging" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658604 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658611 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="northd" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658617 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658626 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="nbdb" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658636 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovn-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658645 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658655 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="kube-rbac-proxy-node" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658787 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658796 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: E1203 12:51:33.658812 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658819 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.658920 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerName="ovnkube-controller" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.660750 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728348 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-var-lib-openvswitch\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728465 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-env-overrides\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728508 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-systemd\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728533 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-node-log\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728546 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728569 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-script-lib\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728746 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-node-log" (OuterVolumeSpecName: "node-log") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.729025 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.729193 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.728738 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42qbm\" (UniqueName: \"kubernetes.io/projected/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-kube-api-access-42qbm\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730114 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-netns\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730156 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-slash\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730183 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-ovn\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730208 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovn-node-metrics-cert\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730239 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-ovn-kubernetes\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730240 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730266 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-netd\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730265 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-slash" (OuterVolumeSpecName: "host-slash") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730284 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730307 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-kubelet\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730325 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730331 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-log-socket\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730359 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-log-socket" (OuterVolumeSpecName: "log-socket") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730346 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730388 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-bin\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730410 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730409 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730439 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-etc-openvswitch\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730487 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730504 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-systemd-units\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730531 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730547 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-config\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730602 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730627 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-openvswitch\") pod \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\" (UID: \"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1\") " Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730692 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.730809 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731043 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731240 4990 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731272 4990 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731295 4990 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731310 4990 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731323 4990 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731334 4990 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731347 4990 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731357 4990 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731367 4990 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731375 4990 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731384 4990 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731393 4990 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731403 4990 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731412 4990 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731424 4990 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731435 4990 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.731446 4990 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.736736 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-kube-api-access-42qbm" (OuterVolumeSpecName: "kube-api-access-42qbm") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "kube-api-access-42qbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.737209 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.747060 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" (UID: "f5d4bd47-2fa6-4e6a-9697-9af4e227afe1"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.832708 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-run-netns\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.839634 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-var-lib-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.839672 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-ovn\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.839703 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-env-overrides\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.839726 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-etc-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.839745 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-run-ovn-kubernetes\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.839775 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840034 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840083 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ftnh\" (UniqueName: \"kubernetes.io/projected/59149aca-b38b-4b58-ac2e-dc54de667383-kube-api-access-6ftnh\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840118 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-cni-netd\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840166 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-systemd\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840218 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-systemd-units\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840378 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-kubelet\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840436 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-node-log\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840486 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59149aca-b38b-4b58-ac2e-dc54de667383-ovn-node-metrics-cert\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840520 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-log-socket\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840591 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-cni-bin\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840642 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-ovnkube-script-lib\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840716 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-slash\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840767 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-ovnkube-config\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840860 4990 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840876 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42qbm\" (UniqueName: \"kubernetes.io/projected/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-kube-api-access-42qbm\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.840896 4990 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942594 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-run-netns\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942678 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-var-lib-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942723 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-ovn\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942763 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-env-overrides\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942792 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-etc-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942778 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-run-netns\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942880 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-var-lib-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942830 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-run-ovn-kubernetes\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.942912 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-run-ovn-kubernetes\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943187 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-etc-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943125 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-ovn\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943265 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943342 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943375 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ftnh\" (UniqueName: \"kubernetes.io/projected/59149aca-b38b-4b58-ac2e-dc54de667383-kube-api-access-6ftnh\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943376 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-openvswitch\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943411 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-cni-netd\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943403 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943497 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-systemd\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943576 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-systemd-units\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943597 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-cni-netd\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943662 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-run-systemd\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943713 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-systemd-units\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943748 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-kubelet\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943803 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-node-log\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943841 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59149aca-b38b-4b58-ac2e-dc54de667383-ovn-node-metrics-cert\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943871 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-kubelet\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943875 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-node-log\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943891 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-log-socket\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943928 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-log-socket\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943953 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-cni-bin\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.943979 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-ovnkube-script-lib\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.944023 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-slash\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.944058 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-ovnkube-config\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.944121 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-env-overrides\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.944985 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-ovnkube-config\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.945025 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-slash\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.945050 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59149aca-b38b-4b58-ac2e-dc54de667383-host-cni-bin\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.945108 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59149aca-b38b-4b58-ac2e-dc54de667383-ovnkube-script-lib\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.959836 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59149aca-b38b-4b58-ac2e-dc54de667383-ovn-node-metrics-cert\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.963350 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ftnh\" (UniqueName: \"kubernetes.io/projected/59149aca-b38b-4b58-ac2e-dc54de667383-kube-api-access-6ftnh\") pod \"ovnkube-node-hnlgs\" (UID: \"59149aca-b38b-4b58-ac2e-dc54de667383\") " pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:33 crc kubenswrapper[4990]: I1203 12:51:33.984086 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:34 crc kubenswrapper[4990]: W1203 12:51:34.007064 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59149aca_b38b_4b58_ac2e_dc54de667383.slice/crio-3182e5ae940d162790523e0d645c57a31a28767ca07e73844893f0d8084da6dc WatchSource:0}: Error finding container 3182e5ae940d162790523e0d645c57a31a28767ca07e73844893f0d8084da6dc: Status 404 returned error can't find the container with id 3182e5ae940d162790523e0d645c57a31a28767ca07e73844893f0d8084da6dc Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.507713 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovn-acl-logging/0.log" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.508519 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7lrq6_f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/ovn-controller/0.log" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.508974 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20" exitCode=0 Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.509003 4990 generic.go:334] "Generic (PLEG): container finished" podID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" containerID="387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9" exitCode=0 Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.509019 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20"} Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.509094 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9"} Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.509131 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" event={"ID":"f5d4bd47-2fa6-4e6a-9697-9af4e227afe1","Type":"ContainerDied","Data":"85ee693ad9ce3c08979c21b865d8709321d746c5acc57d95e081c2fa42b9b3f8"} Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.509159 4990 scope.go:117] "RemoveContainer" containerID="06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.509179 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7lrq6" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.513076 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9c857_41a42811-350f-4311-bb49-f7f116aa6952/kube-multus/2.log" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.513319 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9c857" event={"ID":"41a42811-350f-4311-bb49-f7f116aa6952","Type":"ContainerStarted","Data":"34cf404259d347db8a8e7879eb1c9e3f2ec305829caf4ecaa226c841c7660888"} Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.516527 4990 generic.go:334] "Generic (PLEG): container finished" podID="59149aca-b38b-4b58-ac2e-dc54de667383" containerID="da73d2eae85afa0a83c7c0bd741dca92f39efc0f6b858e24325d947a67c16805" exitCode=0 Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.516608 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerDied","Data":"da73d2eae85afa0a83c7c0bd741dca92f39efc0f6b858e24325d947a67c16805"} Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.516666 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"3182e5ae940d162790523e0d645c57a31a28767ca07e73844893f0d8084da6dc"} Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.530834 4990 scope.go:117] "RemoveContainer" containerID="10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.556958 4990 scope.go:117] "RemoveContainer" containerID="00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.563743 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7lrq6"] Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.570946 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7lrq6"] Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.591168 4990 scope.go:117] "RemoveContainer" containerID="387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.625861 4990 scope.go:117] "RemoveContainer" containerID="b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.656959 4990 scope.go:117] "RemoveContainer" containerID="c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.673494 4990 scope.go:117] "RemoveContainer" containerID="cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.689734 4990 scope.go:117] "RemoveContainer" containerID="122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.710368 4990 scope.go:117] "RemoveContainer" containerID="f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.734443 4990 scope.go:117] "RemoveContainer" containerID="06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.735566 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902\": container with ID starting with 06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902 not found: ID does not exist" containerID="06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.735605 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902"} err="failed to get container status \"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902\": rpc error: code = NotFound desc = could not find container \"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902\": container with ID starting with 06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.735632 4990 scope.go:117] "RemoveContainer" containerID="10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.736226 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\": container with ID starting with 10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729 not found: ID does not exist" containerID="10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.736252 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729"} err="failed to get container status \"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\": rpc error: code = NotFound desc = could not find container \"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\": container with ID starting with 10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.736267 4990 scope.go:117] "RemoveContainer" containerID="00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.736543 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\": container with ID starting with 00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20 not found: ID does not exist" containerID="00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.736560 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20"} err="failed to get container status \"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\": rpc error: code = NotFound desc = could not find container \"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\": container with ID starting with 00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.736574 4990 scope.go:117] "RemoveContainer" containerID="387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.736814 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\": container with ID starting with 387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9 not found: ID does not exist" containerID="387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.736834 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9"} err="failed to get container status \"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\": rpc error: code = NotFound desc = could not find container \"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\": container with ID starting with 387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.736847 4990 scope.go:117] "RemoveContainer" containerID="b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.737120 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\": container with ID starting with b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b not found: ID does not exist" containerID="b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.737144 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b"} err="failed to get container status \"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\": rpc error: code = NotFound desc = could not find container \"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\": container with ID starting with b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.737167 4990 scope.go:117] "RemoveContainer" containerID="c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.737663 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\": container with ID starting with c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d not found: ID does not exist" containerID="c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.737691 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d"} err="failed to get container status \"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\": rpc error: code = NotFound desc = could not find container \"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\": container with ID starting with c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.737711 4990 scope.go:117] "RemoveContainer" containerID="cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.738211 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\": container with ID starting with cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258 not found: ID does not exist" containerID="cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.738232 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258"} err="failed to get container status \"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\": rpc error: code = NotFound desc = could not find container \"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\": container with ID starting with cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.738247 4990 scope.go:117] "RemoveContainer" containerID="122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.738644 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\": container with ID starting with 122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe not found: ID does not exist" containerID="122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.738710 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe"} err="failed to get container status \"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\": rpc error: code = NotFound desc = could not find container \"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\": container with ID starting with 122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.738752 4990 scope.go:117] "RemoveContainer" containerID="f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e" Dec 03 12:51:34 crc kubenswrapper[4990]: E1203 12:51:34.739167 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\": container with ID starting with f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e not found: ID does not exist" containerID="f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.739209 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e"} err="failed to get container status \"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\": rpc error: code = NotFound desc = could not find container \"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\": container with ID starting with f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.739240 4990 scope.go:117] "RemoveContainer" containerID="06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.739654 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902"} err="failed to get container status \"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902\": rpc error: code = NotFound desc = could not find container \"06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902\": container with ID starting with 06b9ed8c27939ff7eeedc0c58c93d3e732ddc375056ae8c0344c1af127629902 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.739697 4990 scope.go:117] "RemoveContainer" containerID="10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.740259 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729"} err="failed to get container status \"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\": rpc error: code = NotFound desc = could not find container \"10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729\": container with ID starting with 10860744cb4fdab88bd8f7eeeb0b0a566324b0a0087d6ec9d98be1060c921729 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.740293 4990 scope.go:117] "RemoveContainer" containerID="00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.740743 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20"} err="failed to get container status \"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\": rpc error: code = NotFound desc = could not find container \"00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20\": container with ID starting with 00935d6763e475a010f98a0b4ce1399d81c298b8a43fce400fceff03e29adc20 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.740765 4990 scope.go:117] "RemoveContainer" containerID="387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.741239 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9"} err="failed to get container status \"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\": rpc error: code = NotFound desc = could not find container \"387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9\": container with ID starting with 387a731c4f6677fba3c62b6e3556e657cc644a445754ae02c2d2c5bbc8dfe0b9 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.741299 4990 scope.go:117] "RemoveContainer" containerID="b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.742118 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b"} err="failed to get container status \"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\": rpc error: code = NotFound desc = could not find container \"b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b\": container with ID starting with b9796385c05aae1fd2b5ddf63f51b12029f373352e0b397932d28aff9a11a09b not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.742157 4990 scope.go:117] "RemoveContainer" containerID="c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.742558 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d"} err="failed to get container status \"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\": rpc error: code = NotFound desc = could not find container \"c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d\": container with ID starting with c1ca3e0d3a5eac2545f7238f605e8fd4b39f238ba3e10e14aeb8c22c7334638d not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.742587 4990 scope.go:117] "RemoveContainer" containerID="cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.742948 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258"} err="failed to get container status \"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\": rpc error: code = NotFound desc = could not find container \"cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258\": container with ID starting with cc3808646538777c9c9b6a0e4db4a8041ccc8b781aafce72dba81c5a68b1b258 not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.743162 4990 scope.go:117] "RemoveContainer" containerID="122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.743902 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe"} err="failed to get container status \"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\": rpc error: code = NotFound desc = could not find container \"122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe\": container with ID starting with 122ef65fe51899ed5a1e44ebff2638cd5b7bf0f65ad6346061f3f1e15238b5fe not found: ID does not exist" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.743961 4990 scope.go:117] "RemoveContainer" containerID="f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e" Dec 03 12:51:34 crc kubenswrapper[4990]: I1203 12:51:34.744703 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e"} err="failed to get container status \"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\": rpc error: code = NotFound desc = could not find container \"f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e\": container with ID starting with f14cb31ac4791c7160f0f05094d317533d0dee35889ece365255d13e52d6c61e not found: ID does not exist" Dec 03 12:51:35 crc kubenswrapper[4990]: I1203 12:51:35.540564 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"d04bdfce4c7fffc13474605229a24d0025b9290775a790d90c4ec6acfa3c60ab"} Dec 03 12:51:35 crc kubenswrapper[4990]: I1203 12:51:35.541435 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"33495df0a3648f35043a221dcd7b6eb1e1145e83523b593647691ad49392df3d"} Dec 03 12:51:35 crc kubenswrapper[4990]: I1203 12:51:35.541469 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"19bbc2e80cf5e86151a44f3241383ab97c04816814d5118f9375acf2d38046bd"} Dec 03 12:51:35 crc kubenswrapper[4990]: I1203 12:51:35.541484 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"324391954e7179115e5180da487a7dfa65378afb604a861d378f504e8bfeb786"} Dec 03 12:51:35 crc kubenswrapper[4990]: I1203 12:51:35.541501 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"f25bf13223e5dba412e06409cb50552728881c43728e13a2c0bbb73e6a739ebb"} Dec 03 12:51:36 crc kubenswrapper[4990]: I1203 12:51:36.276292 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d4bd47-2fa6-4e6a-9697-9af4e227afe1" path="/var/lib/kubelet/pods/f5d4bd47-2fa6-4e6a-9697-9af4e227afe1/volumes" Dec 03 12:51:36 crc kubenswrapper[4990]: I1203 12:51:36.549280 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"c33f1fecf988f212ded3e010bf3abb818323bfc4484565b07438f9f8dd28b9b5"} Dec 03 12:51:38 crc kubenswrapper[4990]: I1203 12:51:38.571404 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"5dfe993afd6b21fbe487dd2b3bfa58583d68bcc996b3a96ee5c9b1daa2453943"} Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.601619 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" event={"ID":"59149aca-b38b-4b58-ac2e-dc54de667383","Type":"ContainerStarted","Data":"ff4ae6ce753ef62762d4377e06c120a3edfbea52a9e25906a1b86489a8bb85e2"} Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.603850 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.603886 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.603949 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.642699 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.642783 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:51:41 crc kubenswrapper[4990]: I1203 12:51:41.651098 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" podStartSLOduration=8.651069199 podStartE2EDuration="8.651069199s" podCreationTimestamp="2025-12-03 12:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:51:41.648349117 +0000 UTC m=+849.790260366" watchObservedRunningTime="2025-12-03 12:51:41.651069199 +0000 UTC m=+849.792980428" Dec 03 12:52:04 crc kubenswrapper[4990]: I1203 12:52:04.007422 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hnlgs" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.149427 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8"] Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.151325 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.153639 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.161659 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8"] Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.311975 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.312090 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xskgm\" (UniqueName: \"kubernetes.io/projected/9c828530-2366-4037-b371-69e9da5a0174-kube-api-access-xskgm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.312127 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.413267 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xskgm\" (UniqueName: \"kubernetes.io/projected/9c828530-2366-4037-b371-69e9da5a0174-kube-api-access-xskgm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.413346 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.413495 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.413871 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.414062 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.440778 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xskgm\" (UniqueName: \"kubernetes.io/projected/9c828530-2366-4037-b371-69e9da5a0174-kube-api-access-xskgm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.464871 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.679518 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8"] Dec 03 12:52:13 crc kubenswrapper[4990]: I1203 12:52:13.775557 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" event={"ID":"9c828530-2366-4037-b371-69e9da5a0174","Type":"ContainerStarted","Data":"a3e4c5f46243e51608f7c68d44ceca013d9f06df6dda4b28d2b12f495884023b"} Dec 03 12:52:14 crc kubenswrapper[4990]: I1203 12:52:14.783077 4990 generic.go:334] "Generic (PLEG): container finished" podID="9c828530-2366-4037-b371-69e9da5a0174" containerID="f8dee73543ae7db05ba663ae85720b3bd7e56e64924b19eae28f429459b9453b" exitCode=0 Dec 03 12:52:14 crc kubenswrapper[4990]: I1203 12:52:14.783402 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" event={"ID":"9c828530-2366-4037-b371-69e9da5a0174","Type":"ContainerDied","Data":"f8dee73543ae7db05ba663ae85720b3bd7e56e64924b19eae28f429459b9453b"} Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.165062 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-44hzr"] Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.166447 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.176017 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44hzr"] Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.339873 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmqwh\" (UniqueName: \"kubernetes.io/projected/8bf6baec-606a-4819-885c-e884e396bd72-kube-api-access-xmqwh\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.339956 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-catalog-content\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.340060 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-utilities\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.441077 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmqwh\" (UniqueName: \"kubernetes.io/projected/8bf6baec-606a-4819-885c-e884e396bd72-kube-api-access-xmqwh\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.441146 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-catalog-content\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.441180 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-utilities\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.441699 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-utilities\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.441752 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-catalog-content\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.470093 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmqwh\" (UniqueName: \"kubernetes.io/projected/8bf6baec-606a-4819-885c-e884e396bd72-kube-api-access-xmqwh\") pod \"redhat-operators-44hzr\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.484554 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.713252 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44hzr"] Dec 03 12:52:15 crc kubenswrapper[4990]: W1203 12:52:15.719061 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bf6baec_606a_4819_885c_e884e396bd72.slice/crio-63dc9357cd814b5f0fbaf029e24ebf619e8ed2a44c7333ef2fe014cb95f3f90d WatchSource:0}: Error finding container 63dc9357cd814b5f0fbaf029e24ebf619e8ed2a44c7333ef2fe014cb95f3f90d: Status 404 returned error can't find the container with id 63dc9357cd814b5f0fbaf029e24ebf619e8ed2a44c7333ef2fe014cb95f3f90d Dec 03 12:52:15 crc kubenswrapper[4990]: I1203 12:52:15.792198 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerStarted","Data":"63dc9357cd814b5f0fbaf029e24ebf619e8ed2a44c7333ef2fe014cb95f3f90d"} Dec 03 12:52:16 crc kubenswrapper[4990]: I1203 12:52:16.798083 4990 generic.go:334] "Generic (PLEG): container finished" podID="8bf6baec-606a-4819-885c-e884e396bd72" containerID="f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562" exitCode=0 Dec 03 12:52:16 crc kubenswrapper[4990]: I1203 12:52:16.798157 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerDied","Data":"f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562"} Dec 03 12:52:16 crc kubenswrapper[4990]: I1203 12:52:16.801515 4990 generic.go:334] "Generic (PLEG): container finished" podID="9c828530-2366-4037-b371-69e9da5a0174" containerID="08ce0da335d5752d6427f683cf4c5d38cbbf3e645fa400bf73ef3b6c71e5640b" exitCode=0 Dec 03 12:52:16 crc kubenswrapper[4990]: I1203 12:52:16.801548 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" event={"ID":"9c828530-2366-4037-b371-69e9da5a0174","Type":"ContainerDied","Data":"08ce0da335d5752d6427f683cf4c5d38cbbf3e645fa400bf73ef3b6c71e5640b"} Dec 03 12:52:17 crc kubenswrapper[4990]: I1203 12:52:17.809207 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerStarted","Data":"40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1"} Dec 03 12:52:17 crc kubenswrapper[4990]: I1203 12:52:17.811139 4990 generic.go:334] "Generic (PLEG): container finished" podID="9c828530-2366-4037-b371-69e9da5a0174" containerID="f92f8918c474fa7abe3c1a8edf6b97343bfee02abb7a2ce96283ebb6b0103cd9" exitCode=0 Dec 03 12:52:17 crc kubenswrapper[4990]: I1203 12:52:17.811172 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" event={"ID":"9c828530-2366-4037-b371-69e9da5a0174","Type":"ContainerDied","Data":"f92f8918c474fa7abe3c1a8edf6b97343bfee02abb7a2ce96283ebb6b0103cd9"} Dec 03 12:52:18 crc kubenswrapper[4990]: I1203 12:52:18.820268 4990 generic.go:334] "Generic (PLEG): container finished" podID="8bf6baec-606a-4819-885c-e884e396bd72" containerID="40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1" exitCode=0 Dec 03 12:52:18 crc kubenswrapper[4990]: I1203 12:52:18.820859 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerDied","Data":"40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1"} Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.134310 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.288239 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xskgm\" (UniqueName: \"kubernetes.io/projected/9c828530-2366-4037-b371-69e9da5a0174-kube-api-access-xskgm\") pod \"9c828530-2366-4037-b371-69e9da5a0174\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.288607 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-util\") pod \"9c828530-2366-4037-b371-69e9da5a0174\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.288679 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-bundle\") pod \"9c828530-2366-4037-b371-69e9da5a0174\" (UID: \"9c828530-2366-4037-b371-69e9da5a0174\") " Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.289362 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-bundle" (OuterVolumeSpecName: "bundle") pod "9c828530-2366-4037-b371-69e9da5a0174" (UID: "9c828530-2366-4037-b371-69e9da5a0174"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.294982 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c828530-2366-4037-b371-69e9da5a0174-kube-api-access-xskgm" (OuterVolumeSpecName: "kube-api-access-xskgm") pod "9c828530-2366-4037-b371-69e9da5a0174" (UID: "9c828530-2366-4037-b371-69e9da5a0174"). InnerVolumeSpecName "kube-api-access-xskgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.390187 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xskgm\" (UniqueName: \"kubernetes.io/projected/9c828530-2366-4037-b371-69e9da5a0174-kube-api-access-xskgm\") on node \"crc\" DevicePath \"\"" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.390222 4990 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.655630 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-util" (OuterVolumeSpecName: "util") pod "9c828530-2366-4037-b371-69e9da5a0174" (UID: "9c828530-2366-4037-b371-69e9da5a0174"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.694243 4990 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9c828530-2366-4037-b371-69e9da5a0174-util\") on node \"crc\" DevicePath \"\"" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.829865 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerStarted","Data":"229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a"} Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.833615 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" event={"ID":"9c828530-2366-4037-b371-69e9da5a0174","Type":"ContainerDied","Data":"a3e4c5f46243e51608f7c68d44ceca013d9f06df6dda4b28d2b12f495884023b"} Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.833812 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3e4c5f46243e51608f7c68d44ceca013d9f06df6dda4b28d2b12f495884023b" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.833677 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8" Dec 03 12:52:19 crc kubenswrapper[4990]: I1203 12:52:19.849461 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-44hzr" podStartSLOduration=1.963421817 podStartE2EDuration="4.849427589s" podCreationTimestamp="2025-12-03 12:52:15 +0000 UTC" firstStartedPulling="2025-12-03 12:52:16.800949601 +0000 UTC m=+884.942860830" lastFinishedPulling="2025-12-03 12:52:19.686955363 +0000 UTC m=+887.828866602" observedRunningTime="2025-12-03 12:52:19.84449585 +0000 UTC m=+887.986407099" watchObservedRunningTime="2025-12-03 12:52:19.849427589 +0000 UTC m=+887.991338818" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.718856 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt"] Dec 03 12:52:23 crc kubenswrapper[4990]: E1203 12:52:23.719361 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="extract" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.719374 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="extract" Dec 03 12:52:23 crc kubenswrapper[4990]: E1203 12:52:23.719385 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="pull" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.719391 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="pull" Dec 03 12:52:23 crc kubenswrapper[4990]: E1203 12:52:23.719411 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="util" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.719418 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="util" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.719554 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c828530-2366-4037-b371-69e9da5a0174" containerName="extract" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.720059 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.722872 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.723010 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-84svp" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.724071 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.735673 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt"] Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.860247 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5xkm\" (UniqueName: \"kubernetes.io/projected/9429610d-e1cb-4248-8224-12e08acf1a69-kube-api-access-c5xkm\") pod \"nmstate-operator-5b5b58f5c8-g69mt\" (UID: \"9429610d-e1cb-4248-8224-12e08acf1a69\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.961262 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5xkm\" (UniqueName: \"kubernetes.io/projected/9429610d-e1cb-4248-8224-12e08acf1a69-kube-api-access-c5xkm\") pod \"nmstate-operator-5b5b58f5c8-g69mt\" (UID: \"9429610d-e1cb-4248-8224-12e08acf1a69\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" Dec 03 12:52:23 crc kubenswrapper[4990]: I1203 12:52:23.987523 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5xkm\" (UniqueName: \"kubernetes.io/projected/9429610d-e1cb-4248-8224-12e08acf1a69-kube-api-access-c5xkm\") pod \"nmstate-operator-5b5b58f5c8-g69mt\" (UID: \"9429610d-e1cb-4248-8224-12e08acf1a69\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" Dec 03 12:52:24 crc kubenswrapper[4990]: I1203 12:52:24.036977 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" Dec 03 12:52:24 crc kubenswrapper[4990]: I1203 12:52:24.500070 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt"] Dec 03 12:52:24 crc kubenswrapper[4990]: I1203 12:52:24.857414 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" event={"ID":"9429610d-e1cb-4248-8224-12e08acf1a69","Type":"ContainerStarted","Data":"3c517d94abdd479d025d5c7646f4a6805418c043f23ac5cd99551895fb0b8b3e"} Dec 03 12:52:25 crc kubenswrapper[4990]: I1203 12:52:25.485121 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:25 crc kubenswrapper[4990]: I1203 12:52:25.485608 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:25 crc kubenswrapper[4990]: I1203 12:52:25.528553 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:25 crc kubenswrapper[4990]: I1203 12:52:25.900399 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:27 crc kubenswrapper[4990]: I1203 12:52:27.755128 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44hzr"] Dec 03 12:52:28 crc kubenswrapper[4990]: I1203 12:52:28.879072 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" event={"ID":"9429610d-e1cb-4248-8224-12e08acf1a69","Type":"ContainerStarted","Data":"832f9cfcc3a700cf4379ef0f1aa3d07346b61ee9de6567da5aca8309b20a015c"} Dec 03 12:52:28 crc kubenswrapper[4990]: I1203 12:52:28.879261 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-44hzr" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="registry-server" containerID="cri-o://229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a" gracePeriod=2 Dec 03 12:52:28 crc kubenswrapper[4990]: I1203 12:52:28.908572 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-g69mt" podStartSLOduration=1.835092796 podStartE2EDuration="5.908552148s" podCreationTimestamp="2025-12-03 12:52:23 +0000 UTC" firstStartedPulling="2025-12-03 12:52:24.515543336 +0000 UTC m=+892.657454565" lastFinishedPulling="2025-12-03 12:52:28.589002688 +0000 UTC m=+896.730913917" observedRunningTime="2025-12-03 12:52:28.901073011 +0000 UTC m=+897.042984250" watchObservedRunningTime="2025-12-03 12:52:28.908552148 +0000 UTC m=+897.050463407" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.761254 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.851191 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-catalog-content\") pod \"8bf6baec-606a-4819-885c-e884e396bd72\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.851322 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-utilities\") pod \"8bf6baec-606a-4819-885c-e884e396bd72\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.851434 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmqwh\" (UniqueName: \"kubernetes.io/projected/8bf6baec-606a-4819-885c-e884e396bd72-kube-api-access-xmqwh\") pod \"8bf6baec-606a-4819-885c-e884e396bd72\" (UID: \"8bf6baec-606a-4819-885c-e884e396bd72\") " Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.852348 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-utilities" (OuterVolumeSpecName: "utilities") pod "8bf6baec-606a-4819-885c-e884e396bd72" (UID: "8bf6baec-606a-4819-885c-e884e396bd72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.857691 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf6baec-606a-4819-885c-e884e396bd72-kube-api-access-xmqwh" (OuterVolumeSpecName: "kube-api-access-xmqwh") pod "8bf6baec-606a-4819-885c-e884e396bd72" (UID: "8bf6baec-606a-4819-885c-e884e396bd72"). InnerVolumeSpecName "kube-api-access-xmqwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.887112 4990 generic.go:334] "Generic (PLEG): container finished" podID="8bf6baec-606a-4819-885c-e884e396bd72" containerID="229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a" exitCode=0 Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.887192 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44hzr" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.887219 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerDied","Data":"229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a"} Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.887296 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44hzr" event={"ID":"8bf6baec-606a-4819-885c-e884e396bd72","Type":"ContainerDied","Data":"63dc9357cd814b5f0fbaf029e24ebf619e8ed2a44c7333ef2fe014cb95f3f90d"} Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.887323 4990 scope.go:117] "RemoveContainer" containerID="229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.912460 4990 scope.go:117] "RemoveContainer" containerID="40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.930809 4990 scope.go:117] "RemoveContainer" containerID="f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.952661 4990 scope.go:117] "RemoveContainer" containerID="229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.953174 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.953213 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmqwh\" (UniqueName: \"kubernetes.io/projected/8bf6baec-606a-4819-885c-e884e396bd72-kube-api-access-xmqwh\") on node \"crc\" DevicePath \"\"" Dec 03 12:52:29 crc kubenswrapper[4990]: E1203 12:52:29.953226 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a\": container with ID starting with 229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a not found: ID does not exist" containerID="229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.953250 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a"} err="failed to get container status \"229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a\": rpc error: code = NotFound desc = could not find container \"229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a\": container with ID starting with 229743f351628413afcd86866d496ee116c56ba0afed90e338fa24c866664a1a not found: ID does not exist" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.953275 4990 scope.go:117] "RemoveContainer" containerID="40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1" Dec 03 12:52:29 crc kubenswrapper[4990]: E1203 12:52:29.960949 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1\": container with ID starting with 40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1 not found: ID does not exist" containerID="40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.961004 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1"} err="failed to get container status \"40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1\": rpc error: code = NotFound desc = could not find container \"40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1\": container with ID starting with 40cff72add6936222c8c22ebeafa2a7c77c8e1d5535ec53a545db04d33f32bb1 not found: ID does not exist" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.961041 4990 scope.go:117] "RemoveContainer" containerID="f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562" Dec 03 12:52:29 crc kubenswrapper[4990]: E1203 12:52:29.961442 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562\": container with ID starting with f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562 not found: ID does not exist" containerID="f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.961491 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562"} err="failed to get container status \"f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562\": rpc error: code = NotFound desc = could not find container \"f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562\": container with ID starting with f09996e6a2b4931201482d0edb63b7bc1d9a6782e93dd28d29bfc5002010f562 not found: ID does not exist" Dec 03 12:52:29 crc kubenswrapper[4990]: I1203 12:52:29.963819 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bf6baec-606a-4819-885c-e884e396bd72" (UID: "8bf6baec-606a-4819-885c-e884e396bd72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:52:30 crc kubenswrapper[4990]: I1203 12:52:30.054068 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf6baec-606a-4819-885c-e884e396bd72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:52:30 crc kubenswrapper[4990]: I1203 12:52:30.226102 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44hzr"] Dec 03 12:52:30 crc kubenswrapper[4990]: I1203 12:52:30.236659 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-44hzr"] Dec 03 12:52:30 crc kubenswrapper[4990]: I1203 12:52:30.271519 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf6baec-606a-4819-885c-e884e396bd72" path="/var/lib/kubelet/pods/8bf6baec-606a-4819-885c-e884e396bd72/volumes" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.286208 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.286594 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.467473 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg"] Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.467719 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="extract-utilities" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.467741 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="extract-utilities" Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.467760 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="extract-content" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.467769 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="extract-content" Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.467787 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="registry-server" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.467794 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="registry-server" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.467934 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf6baec-606a-4819-885c-e884e396bd72" containerName="registry-server" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.468693 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.473788 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-p6rcn" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.494585 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.495394 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.499593 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.530762 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.571579 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-dtt26"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.588842 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.602053 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66xk2\" (UniqueName: \"kubernetes.io/projected/afaa9abc-cdbc-44b9-ab3e-a69df533460c-kube-api-access-66xk2\") pod \"nmstate-metrics-7f946cbc9-gwfbg\" (UID: \"afaa9abc-cdbc-44b9-ab3e-a69df533460c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.602127 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrpt6\" (UniqueName: \"kubernetes.io/projected/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-kube-api-access-mrpt6\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.602174 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.610572 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703262 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-ovs-socket\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703799 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703827 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-nmstate-lock\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703879 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66xk2\" (UniqueName: \"kubernetes.io/projected/afaa9abc-cdbc-44b9-ab3e-a69df533460c-kube-api-access-66xk2\") pod \"nmstate-metrics-7f946cbc9-gwfbg\" (UID: \"afaa9abc-cdbc-44b9-ab3e-a69df533460c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703918 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ftxw\" (UniqueName: \"kubernetes.io/projected/20a68e4a-3638-430c-a27b-7d535613c3d3-kube-api-access-7ftxw\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703955 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrpt6\" (UniqueName: \"kubernetes.io/projected/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-kube-api-access-mrpt6\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.703983 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-dbus-socket\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.704154 4990 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.704212 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-tls-key-pair podName:45d5180b-e7cb-481e-88c7-9fa11bf8edc9 nodeName:}" failed. No retries permitted until 2025-12-03 12:52:34.204191054 +0000 UTC m=+902.346102283 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-pgjpk" (UID: "45d5180b-e7cb-481e-88c7-9fa11bf8edc9") : secret "openshift-nmstate-webhook" not found Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.724233 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.725014 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.731950 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9vcjz" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.731958 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.733006 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.737844 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrpt6\" (UniqueName: \"kubernetes.io/projected/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-kube-api-access-mrpt6\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.739156 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66xk2\" (UniqueName: \"kubernetes.io/projected/afaa9abc-cdbc-44b9-ab3e-a69df533460c-kube-api-access-66xk2\") pod \"nmstate-metrics-7f946cbc9-gwfbg\" (UID: \"afaa9abc-cdbc-44b9-ab3e-a69df533460c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.772539 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.784111 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804714 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-nmstate-lock\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804779 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c717132b-5c00-4d42-9d18-c163e82b27ff-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804838 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c717132b-5c00-4d42-9d18-c163e82b27ff-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804855 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ftxw\" (UniqueName: \"kubernetes.io/projected/20a68e4a-3638-430c-a27b-7d535613c3d3-kube-api-access-7ftxw\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804883 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-dbus-socket\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804898 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-ovs-socket\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.804946 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng9gd\" (UniqueName: \"kubernetes.io/projected/c717132b-5c00-4d42-9d18-c163e82b27ff-kube-api-access-ng9gd\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.805132 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-nmstate-lock\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.805482 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-ovs-socket\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.805514 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/20a68e4a-3638-430c-a27b-7d535613c3d3-dbus-socket\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.829559 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ftxw\" (UniqueName: \"kubernetes.io/projected/20a68e4a-3638-430c-a27b-7d535613c3d3-kube-api-access-7ftxw\") pod \"nmstate-handler-dtt26\" (UID: \"20a68e4a-3638-430c-a27b-7d535613c3d3\") " pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.906481 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c717132b-5c00-4d42-9d18-c163e82b27ff-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.906680 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c717132b-5c00-4d42-9d18-c163e82b27ff-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.906748 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng9gd\" (UniqueName: \"kubernetes.io/projected/c717132b-5c00-4d42-9d18-c163e82b27ff-kube-api-access-ng9gd\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.906849 4990 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 03 12:52:33 crc kubenswrapper[4990]: E1203 12:52:33.906918 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c717132b-5c00-4d42-9d18-c163e82b27ff-plugin-serving-cert podName:c717132b-5c00-4d42-9d18-c163e82b27ff nodeName:}" failed. No retries permitted until 2025-12-03 12:52:34.406901592 +0000 UTC m=+902.548812821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/c717132b-5c00-4d42-9d18-c163e82b27ff-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-7n2ph" (UID: "c717132b-5c00-4d42-9d18-c163e82b27ff") : secret "plugin-serving-cert" not found Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.909889 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c717132b-5c00-4d42-9d18-c163e82b27ff-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.930976 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.931870 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng9gd\" (UniqueName: \"kubernetes.io/projected/c717132b-5c00-4d42-9d18-c163e82b27ff-kube-api-access-ng9gd\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.972949 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-77c9cb7494-sb99d"] Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.973729 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:33 crc kubenswrapper[4990]: I1203 12:52:33.978888 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77c9cb7494-sb99d"] Dec 03 12:52:33 crc kubenswrapper[4990]: W1203 12:52:33.985282 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20a68e4a_3638_430c_a27b_7d535613c3d3.slice/crio-b5f730932aea56bb02484803b8a2b02ce2fe7bdeace0aad625ee9386d79d7986 WatchSource:0}: Error finding container b5f730932aea56bb02484803b8a2b02ce2fe7bdeace0aad625ee9386d79d7986: Status 404 returned error can't find the container with id b5f730932aea56bb02484803b8a2b02ce2fe7bdeace0aad625ee9386d79d7986 Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.059133 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg"] Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112211 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-oauth-serving-cert\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112314 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-oauth-config\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112348 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-trusted-ca-bundle\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112386 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-config\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112411 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-serving-cert\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112484 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4lcz\" (UniqueName: \"kubernetes.io/projected/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-kube-api-access-t4lcz\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.112528 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-service-ca\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213578 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4lcz\" (UniqueName: \"kubernetes.io/projected/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-kube-api-access-t4lcz\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213681 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-service-ca\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213734 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-oauth-serving-cert\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213770 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213815 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-oauth-config\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213850 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-trusted-ca-bundle\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213882 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-config\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.213907 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-serving-cert\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.214973 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-service-ca\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.215743 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-trusted-ca-bundle\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.215837 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-oauth-serving-cert\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.215765 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-config\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.217539 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-oauth-config\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.218000 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/45d5180b-e7cb-481e-88c7-9fa11bf8edc9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pgjpk\" (UID: \"45d5180b-e7cb-481e-88c7-9fa11bf8edc9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.218484 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-console-serving-cert\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.233262 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4lcz\" (UniqueName: \"kubernetes.io/projected/aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f-kube-api-access-t4lcz\") pod \"console-77c9cb7494-sb99d\" (UID: \"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f\") " pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.315288 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.412847 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.416862 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c717132b-5c00-4d42-9d18-c163e82b27ff-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.420816 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c717132b-5c00-4d42-9d18-c163e82b27ff-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7n2ph\" (UID: \"c717132b-5c00-4d42-9d18-c163e82b27ff\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.478407 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.542784 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77c9cb7494-sb99d"] Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.627656 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk"] Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.695818 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph"] Dec 03 12:52:34 crc kubenswrapper[4990]: W1203 12:52:34.700136 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc717132b_5c00_4d42_9d18_c163e82b27ff.slice/crio-dc509791595552646cc818b643a6ad098b98a83a616db3052c8cc0db8767fdb5 WatchSource:0}: Error finding container dc509791595552646cc818b643a6ad098b98a83a616db3052c8cc0db8767fdb5: Status 404 returned error can't find the container with id dc509791595552646cc818b643a6ad098b98a83a616db3052c8cc0db8767fdb5 Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.918820 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dtt26" event={"ID":"20a68e4a-3638-430c-a27b-7d535613c3d3","Type":"ContainerStarted","Data":"b5f730932aea56bb02484803b8a2b02ce2fe7bdeace0aad625ee9386d79d7986"} Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.920385 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77c9cb7494-sb99d" event={"ID":"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f","Type":"ContainerStarted","Data":"b3a6e79eeb2e35645937e967a7b654eca190d3c2106fa13b751b43d830e0d660"} Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.920440 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77c9cb7494-sb99d" event={"ID":"aafad52d-47e0-4a7b-a50b-bc0bc8f7a06f","Type":"ContainerStarted","Data":"bb92d411e07598970af8ba2a4618f25e748e298c8606c7eca1199e817c2233e1"} Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.922034 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" event={"ID":"afaa9abc-cdbc-44b9-ab3e-a69df533460c","Type":"ContainerStarted","Data":"2ed270b9e58f24b5709dded93adba326572a86be285eb1a3e9f2d3c56058b5ed"} Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.923342 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" event={"ID":"c717132b-5c00-4d42-9d18-c163e82b27ff","Type":"ContainerStarted","Data":"dc509791595552646cc818b643a6ad098b98a83a616db3052c8cc0db8767fdb5"} Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.924735 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" event={"ID":"45d5180b-e7cb-481e-88c7-9fa11bf8edc9","Type":"ContainerStarted","Data":"5cc41107137e91089b29a7f971c8045c05d93a3835e6ff541926a51e27fff15b"} Dec 03 12:52:34 crc kubenswrapper[4990]: I1203 12:52:34.941057 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-77c9cb7494-sb99d" podStartSLOduration=1.9410370559999999 podStartE2EDuration="1.941037056s" podCreationTimestamp="2025-12-03 12:52:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:52:34.936297581 +0000 UTC m=+903.078208810" watchObservedRunningTime="2025-12-03 12:52:34.941037056 +0000 UTC m=+903.082948285" Dec 03 12:52:36 crc kubenswrapper[4990]: I1203 12:52:36.938556 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" event={"ID":"afaa9abc-cdbc-44b9-ab3e-a69df533460c","Type":"ContainerStarted","Data":"d5bbeaef5763bbddffeab46b97941d49ea2a910cdb7ecb1e692ecbf83888e1ab"} Dec 03 12:52:36 crc kubenswrapper[4990]: I1203 12:52:36.940172 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" event={"ID":"45d5180b-e7cb-481e-88c7-9fa11bf8edc9","Type":"ContainerStarted","Data":"fd5db63860a6910aa10b7d0853e557539cda10f82a4a4be6341f0fc7101ecada"} Dec 03 12:52:36 crc kubenswrapper[4990]: I1203 12:52:36.940344 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:52:36 crc kubenswrapper[4990]: I1203 12:52:36.958800 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" podStartSLOduration=2.032381977 podStartE2EDuration="3.958776021s" podCreationTimestamp="2025-12-03 12:52:33 +0000 UTC" firstStartedPulling="2025-12-03 12:52:34.64173215 +0000 UTC m=+902.783643379" lastFinishedPulling="2025-12-03 12:52:36.568126194 +0000 UTC m=+904.710037423" observedRunningTime="2025-12-03 12:52:36.954221361 +0000 UTC m=+905.096132600" watchObservedRunningTime="2025-12-03 12:52:36.958776021 +0000 UTC m=+905.100687270" Dec 03 12:52:37 crc kubenswrapper[4990]: I1203 12:52:37.950871 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dtt26" event={"ID":"20a68e4a-3638-430c-a27b-7d535613c3d3","Type":"ContainerStarted","Data":"7f4486014015abbc001e8b8097d04f6dfc1f482880c5cdc1da238f0ecb8a31ac"} Dec 03 12:52:37 crc kubenswrapper[4990]: I1203 12:52:37.970395 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-dtt26" podStartSLOduration=2.4369912400000002 podStartE2EDuration="4.970377519s" podCreationTimestamp="2025-12-03 12:52:33 +0000 UTC" firstStartedPulling="2025-12-03 12:52:33.991380101 +0000 UTC m=+902.133291330" lastFinishedPulling="2025-12-03 12:52:36.52476638 +0000 UTC m=+904.666677609" observedRunningTime="2025-12-03 12:52:37.967147194 +0000 UTC m=+906.109058433" watchObservedRunningTime="2025-12-03 12:52:37.970377519 +0000 UTC m=+906.112288748" Dec 03 12:52:38 crc kubenswrapper[4990]: I1203 12:52:38.931857 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:38 crc kubenswrapper[4990]: I1203 12:52:38.958118 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" event={"ID":"c717132b-5c00-4d42-9d18-c163e82b27ff","Type":"ContainerStarted","Data":"803811bfd3d2d5e481e75f019e8c01350bdd7046633bd5c6e43f01863e25ee0f"} Dec 03 12:52:39 crc kubenswrapper[4990]: I1203 12:52:39.965620 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" event={"ID":"afaa9abc-cdbc-44b9-ab3e-a69df533460c","Type":"ContainerStarted","Data":"47dfb0aaf491534a12698506c3f1e6a59ede4f5b7db2125d93e44f0b144a5b11"} Dec 03 12:52:39 crc kubenswrapper[4990]: I1203 12:52:39.989670 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-gwfbg" podStartSLOduration=1.53331779 podStartE2EDuration="6.989650725s" podCreationTimestamp="2025-12-03 12:52:33 +0000 UTC" firstStartedPulling="2025-12-03 12:52:34.06827982 +0000 UTC m=+902.210191049" lastFinishedPulling="2025-12-03 12:52:39.524612755 +0000 UTC m=+907.666523984" observedRunningTime="2025-12-03 12:52:39.984483338 +0000 UTC m=+908.126394567" watchObservedRunningTime="2025-12-03 12:52:39.989650725 +0000 UTC m=+908.131561954" Dec 03 12:52:39 crc kubenswrapper[4990]: I1203 12:52:39.990069 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7n2ph" podStartSLOduration=3.629507243 podStartE2EDuration="6.990061055s" podCreationTimestamp="2025-12-03 12:52:33 +0000 UTC" firstStartedPulling="2025-12-03 12:52:34.701889577 +0000 UTC m=+902.843800806" lastFinishedPulling="2025-12-03 12:52:38.062443389 +0000 UTC m=+906.204354618" observedRunningTime="2025-12-03 12:52:38.975704013 +0000 UTC m=+907.117615242" watchObservedRunningTime="2025-12-03 12:52:39.990061055 +0000 UTC m=+908.131972284" Dec 03 12:52:43 crc kubenswrapper[4990]: I1203 12:52:43.953742 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-dtt26" Dec 03 12:52:44 crc kubenswrapper[4990]: I1203 12:52:44.316114 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:44 crc kubenswrapper[4990]: I1203 12:52:44.316412 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:44 crc kubenswrapper[4990]: I1203 12:52:44.321063 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:44 crc kubenswrapper[4990]: I1203 12:52:44.997322 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-77c9cb7494-sb99d" Dec 03 12:52:45 crc kubenswrapper[4990]: I1203 12:52:45.091482 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-59lwv"] Dec 03 12:52:54 crc kubenswrapper[4990]: I1203 12:52:54.420334 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pgjpk" Dec 03 12:53:03 crc kubenswrapper[4990]: I1203 12:53:03.286226 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:53:03 crc kubenswrapper[4990]: I1203 12:53:03.286749 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.360705 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns"] Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.362560 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.364691 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.372326 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns"] Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.438951 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lttbd\" (UniqueName: \"kubernetes.io/projected/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-kube-api-access-lttbd\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.439029 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.439212 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.541038 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.541484 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.541637 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.541648 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lttbd\" (UniqueName: \"kubernetes.io/projected/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-kube-api-access-lttbd\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.542261 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.562622 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lttbd\" (UniqueName: \"kubernetes.io/projected/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-kube-api-access-lttbd\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:08 crc kubenswrapper[4990]: I1203 12:53:08.677782 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:09 crc kubenswrapper[4990]: I1203 12:53:09.134328 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns"] Dec 03 12:53:09 crc kubenswrapper[4990]: W1203 12:53:09.144832 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97bd525e_80e4_4b77_b2b6_e8d75c9ff66f.slice/crio-786c557f8509402306ccd58f00492cd7c5bd025f40c83d0b366f2fa1bfb86ba6 WatchSource:0}: Error finding container 786c557f8509402306ccd58f00492cd7c5bd025f40c83d0b366f2fa1bfb86ba6: Status 404 returned error can't find the container with id 786c557f8509402306ccd58f00492cd7c5bd025f40c83d0b366f2fa1bfb86ba6 Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.142478 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-59lwv" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerName="console" containerID="cri-o://d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe" gracePeriod=15 Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.145767 4990 generic.go:334] "Generic (PLEG): container finished" podID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerID="5c77e06a3aecd6376820131332721fb12616666aedd7566d25a94d9b695ea69e" exitCode=0 Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.145822 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" event={"ID":"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f","Type":"ContainerDied","Data":"5c77e06a3aecd6376820131332721fb12616666aedd7566d25a94d9b695ea69e"} Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.145850 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" event={"ID":"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f","Type":"ContainerStarted","Data":"786c557f8509402306ccd58f00492cd7c5bd025f40c83d0b366f2fa1bfb86ba6"} Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.492296 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-59lwv_cfc3766b-745f-4fe8-b1e5-beb39019ab01/console/0.log" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.492364 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.572945 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-oauth-config\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.573044 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-trusted-ca-bundle\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.573114 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-serving-cert\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.573144 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-oauth-serving-cert\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.573192 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-config\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.573216 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-service-ca\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.573247 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4956j\" (UniqueName: \"kubernetes.io/projected/cfc3766b-745f-4fe8-b1e5-beb39019ab01-kube-api-access-4956j\") pod \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\" (UID: \"cfc3766b-745f-4fe8-b1e5-beb39019ab01\") " Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.574097 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.574131 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.574146 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-config" (OuterVolumeSpecName: "console-config") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.574564 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-service-ca" (OuterVolumeSpecName: "service-ca") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.578994 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc3766b-745f-4fe8-b1e5-beb39019ab01-kube-api-access-4956j" (OuterVolumeSpecName: "kube-api-access-4956j") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "kube-api-access-4956j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.583740 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.583893 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cfc3766b-745f-4fe8-b1e5-beb39019ab01" (UID: "cfc3766b-745f-4fe8-b1e5-beb39019ab01"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674417 4990 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674519 4990 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674528 4990 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674540 4990 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674548 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4956j\" (UniqueName: \"kubernetes.io/projected/cfc3766b-745f-4fe8-b1e5-beb39019ab01-kube-api-access-4956j\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674558 4990 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfc3766b-745f-4fe8-b1e5-beb39019ab01-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:10 crc kubenswrapper[4990]: I1203 12:53:10.674566 4990 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfc3766b-745f-4fe8-b1e5-beb39019ab01-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.153149 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-59lwv_cfc3766b-745f-4fe8-b1e5-beb39019ab01/console/0.log" Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.153197 4990 generic.go:334] "Generic (PLEG): container finished" podID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerID="d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe" exitCode=2 Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.153224 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-59lwv" event={"ID":"cfc3766b-745f-4fe8-b1e5-beb39019ab01","Type":"ContainerDied","Data":"d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe"} Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.153248 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-59lwv" event={"ID":"cfc3766b-745f-4fe8-b1e5-beb39019ab01","Type":"ContainerDied","Data":"a0d3f6667ac90d012b7bfb6f9685d599058c20b06f18a71a3e7f3d14de4e815e"} Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.153253 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-59lwv" Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.153264 4990 scope.go:117] "RemoveContainer" containerID="d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe" Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.184204 4990 scope.go:117] "RemoveContainer" containerID="d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe" Dec 03 12:53:11 crc kubenswrapper[4990]: E1203 12:53:11.184959 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe\": container with ID starting with d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe not found: ID does not exist" containerID="d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe" Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.185016 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe"} err="failed to get container status \"d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe\": rpc error: code = NotFound desc = could not find container \"d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe\": container with ID starting with d4af981e1ca25886091f1e4c0887d9d45a3e04cc68bfd6462183836c8106bffe not found: ID does not exist" Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.185776 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-59lwv"] Dec 03 12:53:11 crc kubenswrapper[4990]: I1203 12:53:11.189382 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-59lwv"] Dec 03 12:53:12 crc kubenswrapper[4990]: I1203 12:53:12.162154 4990 generic.go:334] "Generic (PLEG): container finished" podID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerID="d98665c581a900dd788955923c47549887507c6ac7d5eecfdf9956130f9d7e1b" exitCode=0 Dec 03 12:53:12 crc kubenswrapper[4990]: I1203 12:53:12.162219 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" event={"ID":"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f","Type":"ContainerDied","Data":"d98665c581a900dd788955923c47549887507c6ac7d5eecfdf9956130f9d7e1b"} Dec 03 12:53:12 crc kubenswrapper[4990]: I1203 12:53:12.272913 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" path="/var/lib/kubelet/pods/cfc3766b-745f-4fe8-b1e5-beb39019ab01/volumes" Dec 03 12:53:13 crc kubenswrapper[4990]: I1203 12:53:13.172218 4990 generic.go:334] "Generic (PLEG): container finished" podID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerID="50d990c70e7b9605ce5d6b01314450cc3342b152cdb9f1b044d79633a6d07e88" exitCode=0 Dec 03 12:53:13 crc kubenswrapper[4990]: I1203 12:53:13.172269 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" event={"ID":"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f","Type":"ContainerDied","Data":"50d990c70e7b9605ce5d6b01314450cc3342b152cdb9f1b044d79633a6d07e88"} Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.391151 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.435522 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-util\") pod \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.435649 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lttbd\" (UniqueName: \"kubernetes.io/projected/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-kube-api-access-lttbd\") pod \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.435691 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-bundle\") pod \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\" (UID: \"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f\") " Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.436854 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-bundle" (OuterVolumeSpecName: "bundle") pod "97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" (UID: "97bd525e-80e4-4b77-b2b6-e8d75c9ff66f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.442328 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-kube-api-access-lttbd" (OuterVolumeSpecName: "kube-api-access-lttbd") pod "97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" (UID: "97bd525e-80e4-4b77-b2b6-e8d75c9ff66f"). InnerVolumeSpecName "kube-api-access-lttbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.451862 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-util" (OuterVolumeSpecName: "util") pod "97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" (UID: "97bd525e-80e4-4b77-b2b6-e8d75c9ff66f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.537589 4990 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-util\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.537639 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lttbd\" (UniqueName: \"kubernetes.io/projected/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-kube-api-access-lttbd\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:14 crc kubenswrapper[4990]: I1203 12:53:14.537653 4990 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97bd525e-80e4-4b77-b2b6-e8d75c9ff66f-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:53:15 crc kubenswrapper[4990]: I1203 12:53:15.185339 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" event={"ID":"97bd525e-80e4-4b77-b2b6-e8d75c9ff66f","Type":"ContainerDied","Data":"786c557f8509402306ccd58f00492cd7c5bd025f40c83d0b366f2fa1bfb86ba6"} Dec 03 12:53:15 crc kubenswrapper[4990]: I1203 12:53:15.185384 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="786c557f8509402306ccd58f00492cd7c5bd025f40c83d0b366f2fa1bfb86ba6" Dec 03 12:53:15 crc kubenswrapper[4990]: I1203 12:53:15.185414 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.764260 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-584cf47849-n84tl"] Dec 03 12:53:23 crc kubenswrapper[4990]: E1203 12:53:23.765144 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="pull" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765163 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="pull" Dec 03 12:53:23 crc kubenswrapper[4990]: E1203 12:53:23.765186 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerName="console" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765193 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerName="console" Dec 03 12:53:23 crc kubenswrapper[4990]: E1203 12:53:23.765211 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="util" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765217 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="util" Dec 03 12:53:23 crc kubenswrapper[4990]: E1203 12:53:23.765226 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="extract" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765232 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="extract" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765346 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc3766b-745f-4fe8-b1e5-beb39019ab01" containerName="console" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765369 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="97bd525e-80e4-4b77-b2b6-e8d75c9ff66f" containerName="extract" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.765877 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.768768 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8m57t" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.768890 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.769087 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.770728 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.770938 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.788955 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-584cf47849-n84tl"] Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.861053 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5aac7d67-caa8-481d-80fb-7ca848f1007f-webhook-cert\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.861149 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmdzx\" (UniqueName: \"kubernetes.io/projected/5aac7d67-caa8-481d-80fb-7ca848f1007f-kube-api-access-pmdzx\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.861184 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5aac7d67-caa8-481d-80fb-7ca848f1007f-apiservice-cert\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.962211 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5aac7d67-caa8-481d-80fb-7ca848f1007f-webhook-cert\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.963406 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmdzx\" (UniqueName: \"kubernetes.io/projected/5aac7d67-caa8-481d-80fb-7ca848f1007f-kube-api-access-pmdzx\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.963471 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5aac7d67-caa8-481d-80fb-7ca848f1007f-apiservice-cert\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.970757 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5aac7d67-caa8-481d-80fb-7ca848f1007f-webhook-cert\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.974266 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5aac7d67-caa8-481d-80fb-7ca848f1007f-apiservice-cert\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:23 crc kubenswrapper[4990]: I1203 12:53:23.985624 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmdzx\" (UniqueName: \"kubernetes.io/projected/5aac7d67-caa8-481d-80fb-7ca848f1007f-kube-api-access-pmdzx\") pod \"metallb-operator-controller-manager-584cf47849-n84tl\" (UID: \"5aac7d67-caa8-481d-80fb-7ca848f1007f\") " pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.083899 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.126044 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-8464f887b-s5zch"] Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.127056 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.135298 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2jg97" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.136176 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.136311 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.212192 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8464f887b-s5zch"] Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.271164 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzdqv\" (UniqueName: \"kubernetes.io/projected/1d95a171-2b09-4441-b7e7-31a5f4427a50-kube-api-access-zzdqv\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.271580 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d95a171-2b09-4441-b7e7-31a5f4427a50-apiservice-cert\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.271613 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d95a171-2b09-4441-b7e7-31a5f4427a50-webhook-cert\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.373852 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzdqv\" (UniqueName: \"kubernetes.io/projected/1d95a171-2b09-4441-b7e7-31a5f4427a50-kube-api-access-zzdqv\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.373950 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d95a171-2b09-4441-b7e7-31a5f4427a50-apiservice-cert\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.373977 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d95a171-2b09-4441-b7e7-31a5f4427a50-webhook-cert\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.385245 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d95a171-2b09-4441-b7e7-31a5f4427a50-apiservice-cert\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.386619 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d95a171-2b09-4441-b7e7-31a5f4427a50-webhook-cert\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.407747 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzdqv\" (UniqueName: \"kubernetes.io/projected/1d95a171-2b09-4441-b7e7-31a5f4427a50-kube-api-access-zzdqv\") pod \"metallb-operator-webhook-server-8464f887b-s5zch\" (UID: \"1d95a171-2b09-4441-b7e7-31a5f4427a50\") " pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.461786 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.555034 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-584cf47849-n84tl"] Dec 03 12:53:24 crc kubenswrapper[4990]: I1203 12:53:24.925264 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8464f887b-s5zch"] Dec 03 12:53:24 crc kubenswrapper[4990]: W1203 12:53:24.932550 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d95a171_2b09_4441_b7e7_31a5f4427a50.slice/crio-8f3a6e5b2e76874449fb9445a88a7b7c35bab2818dec1356473c9e967db7f919 WatchSource:0}: Error finding container 8f3a6e5b2e76874449fb9445a88a7b7c35bab2818dec1356473c9e967db7f919: Status 404 returned error can't find the container with id 8f3a6e5b2e76874449fb9445a88a7b7c35bab2818dec1356473c9e967db7f919 Dec 03 12:53:25 crc kubenswrapper[4990]: I1203 12:53:25.240944 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" event={"ID":"5aac7d67-caa8-481d-80fb-7ca848f1007f","Type":"ContainerStarted","Data":"46d0e9571ed538005378a0a334bb272063401d8b6b0e33837e5143f75b6fef19"} Dec 03 12:53:25 crc kubenswrapper[4990]: I1203 12:53:25.243359 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" event={"ID":"1d95a171-2b09-4441-b7e7-31a5f4427a50","Type":"ContainerStarted","Data":"8f3a6e5b2e76874449fb9445a88a7b7c35bab2818dec1356473c9e967db7f919"} Dec 03 12:53:32 crc kubenswrapper[4990]: I1203 12:53:32.291299 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" event={"ID":"5aac7d67-caa8-481d-80fb-7ca848f1007f","Type":"ContainerStarted","Data":"c9a5a8d888b2e65043b043e659004679f85cda57a1d62718d46dbf81a4e144b2"} Dec 03 12:53:32 crc kubenswrapper[4990]: I1203 12:53:32.292250 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:53:32 crc kubenswrapper[4990]: I1203 12:53:32.294121 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" event={"ID":"1d95a171-2b09-4441-b7e7-31a5f4427a50","Type":"ContainerStarted","Data":"ad52a11c505a17098370e0565400ba7faa4e5db84bc17c871d0223a819ae1ae7"} Dec 03 12:53:32 crc kubenswrapper[4990]: I1203 12:53:32.294260 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:32 crc kubenswrapper[4990]: I1203 12:53:32.332289 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" podStartSLOduration=1.9100583150000001 podStartE2EDuration="8.332265631s" podCreationTimestamp="2025-12-03 12:53:24 +0000 UTC" firstStartedPulling="2025-12-03 12:53:24.935855922 +0000 UTC m=+953.077767151" lastFinishedPulling="2025-12-03 12:53:31.358063238 +0000 UTC m=+959.499974467" observedRunningTime="2025-12-03 12:53:32.328615615 +0000 UTC m=+960.470526864" watchObservedRunningTime="2025-12-03 12:53:32.332265631 +0000 UTC m=+960.474176860" Dec 03 12:53:32 crc kubenswrapper[4990]: I1203 12:53:32.356911 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" podStartSLOduration=2.669307252 podStartE2EDuration="9.356890601s" podCreationTimestamp="2025-12-03 12:53:23 +0000 UTC" firstStartedPulling="2025-12-03 12:53:24.651266613 +0000 UTC m=+952.793177842" lastFinishedPulling="2025-12-03 12:53:31.338849962 +0000 UTC m=+959.480761191" observedRunningTime="2025-12-03 12:53:32.354292392 +0000 UTC m=+960.496203621" watchObservedRunningTime="2025-12-03 12:53:32.356890601 +0000 UTC m=+960.498801830" Dec 03 12:53:33 crc kubenswrapper[4990]: I1203 12:53:33.286872 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:53:33 crc kubenswrapper[4990]: I1203 12:53:33.287231 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:53:33 crc kubenswrapper[4990]: I1203 12:53:33.287282 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:53:33 crc kubenswrapper[4990]: I1203 12:53:33.287918 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca4307683ea7333a625e2203775fa26ab0df3b188bf9bac1cd9b6194d3df672f"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 12:53:33 crc kubenswrapper[4990]: I1203 12:53:33.287990 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://ca4307683ea7333a625e2203775fa26ab0df3b188bf9bac1cd9b6194d3df672f" gracePeriod=600 Dec 03 12:53:34 crc kubenswrapper[4990]: I1203 12:53:34.307882 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="ca4307683ea7333a625e2203775fa26ab0df3b188bf9bac1cd9b6194d3df672f" exitCode=0 Dec 03 12:53:34 crc kubenswrapper[4990]: I1203 12:53:34.307959 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"ca4307683ea7333a625e2203775fa26ab0df3b188bf9bac1cd9b6194d3df672f"} Dec 03 12:53:34 crc kubenswrapper[4990]: I1203 12:53:34.308894 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"80c0112fe475b4f4e2b7d1d072f2e686f621ee22a4d45518867b967855b8112b"} Dec 03 12:53:34 crc kubenswrapper[4990]: I1203 12:53:34.308920 4990 scope.go:117] "RemoveContainer" containerID="0e6f8ca4fb394aa93d3f85ed97825d9ed55398588ce9a009c2c47f07ca7524e8" Dec 03 12:53:44 crc kubenswrapper[4990]: I1203 12:53:44.478430 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-8464f887b-s5zch" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.648836 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hvzjl"] Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.650257 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.670632 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hvzjl"] Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.754722 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbxl2\" (UniqueName: \"kubernetes.io/projected/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-kube-api-access-fbxl2\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.755053 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-utilities\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.755190 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-catalog-content\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.857123 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-utilities\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.857212 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-catalog-content\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.857245 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbxl2\" (UniqueName: \"kubernetes.io/projected/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-kube-api-access-fbxl2\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.857973 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-utilities\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.858053 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-catalog-content\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.879409 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbxl2\" (UniqueName: \"kubernetes.io/projected/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-kube-api-access-fbxl2\") pod \"community-operators-hvzjl\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:46 crc kubenswrapper[4990]: I1203 12:53:46.970461 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:47 crc kubenswrapper[4990]: I1203 12:53:47.434538 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hvzjl"] Dec 03 12:53:48 crc kubenswrapper[4990]: I1203 12:53:48.411644 4990 generic.go:334] "Generic (PLEG): container finished" podID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerID="d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c" exitCode=0 Dec 03 12:53:48 crc kubenswrapper[4990]: I1203 12:53:48.411739 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerDied","Data":"d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c"} Dec 03 12:53:48 crc kubenswrapper[4990]: I1203 12:53:48.413089 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerStarted","Data":"17dc8b06cee3c6c1a16b8a0676804cf8f8cd754407e34a0fda47071d4b579aa5"} Dec 03 12:53:49 crc kubenswrapper[4990]: I1203 12:53:49.426038 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerStarted","Data":"399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f"} Dec 03 12:53:50 crc kubenswrapper[4990]: I1203 12:53:50.432610 4990 generic.go:334] "Generic (PLEG): container finished" podID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerID="399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f" exitCode=0 Dec 03 12:53:50 crc kubenswrapper[4990]: I1203 12:53:50.432664 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerDied","Data":"399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f"} Dec 03 12:53:52 crc kubenswrapper[4990]: I1203 12:53:52.448171 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerStarted","Data":"37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e"} Dec 03 12:53:52 crc kubenswrapper[4990]: I1203 12:53:52.468527 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hvzjl" podStartSLOduration=2.806655155 podStartE2EDuration="6.468506819s" podCreationTimestamp="2025-12-03 12:53:46 +0000 UTC" firstStartedPulling="2025-12-03 12:53:48.413134839 +0000 UTC m=+976.555046078" lastFinishedPulling="2025-12-03 12:53:52.074986513 +0000 UTC m=+980.216897742" observedRunningTime="2025-12-03 12:53:52.466359632 +0000 UTC m=+980.608270861" watchObservedRunningTime="2025-12-03 12:53:52.468506819 +0000 UTC m=+980.610418088" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.637278 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f7w54"] Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.638786 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.650698 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f7w54"] Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.676253 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5nsq\" (UniqueName: \"kubernetes.io/projected/c8dc90fc-14d0-4d60-8cff-6b58658252e0-kube-api-access-n5nsq\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.676358 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-utilities\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.676403 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-catalog-content\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.777570 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-utilities\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.777634 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-catalog-content\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.777666 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5nsq\" (UniqueName: \"kubernetes.io/projected/c8dc90fc-14d0-4d60-8cff-6b58658252e0-kube-api-access-n5nsq\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.778270 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-catalog-content\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.778307 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-utilities\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.806472 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5nsq\" (UniqueName: \"kubernetes.io/projected/c8dc90fc-14d0-4d60-8cff-6b58658252e0-kube-api-access-n5nsq\") pod \"certified-operators-f7w54\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:55 crc kubenswrapper[4990]: I1203 12:53:55.959979 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:53:56 crc kubenswrapper[4990]: I1203 12:53:56.405793 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f7w54"] Dec 03 12:53:56 crc kubenswrapper[4990]: W1203 12:53:56.410934 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8dc90fc_14d0_4d60_8cff_6b58658252e0.slice/crio-4d6063bd8f02aacf40330207c62f019c43e13fd6271711d36959d307468b09c7 WatchSource:0}: Error finding container 4d6063bd8f02aacf40330207c62f019c43e13fd6271711d36959d307468b09c7: Status 404 returned error can't find the container with id 4d6063bd8f02aacf40330207c62f019c43e13fd6271711d36959d307468b09c7 Dec 03 12:53:56 crc kubenswrapper[4990]: I1203 12:53:56.480411 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerStarted","Data":"4d6063bd8f02aacf40330207c62f019c43e13fd6271711d36959d307468b09c7"} Dec 03 12:53:56 crc kubenswrapper[4990]: I1203 12:53:56.971383 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:56 crc kubenswrapper[4990]: I1203 12:53:56.971437 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:57 crc kubenswrapper[4990]: I1203 12:53:57.012806 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:57 crc kubenswrapper[4990]: I1203 12:53:57.487088 4990 generic.go:334] "Generic (PLEG): container finished" podID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerID="e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632" exitCode=0 Dec 03 12:53:57 crc kubenswrapper[4990]: I1203 12:53:57.487140 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerDied","Data":"e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632"} Dec 03 12:53:57 crc kubenswrapper[4990]: I1203 12:53:57.531804 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:53:58 crc kubenswrapper[4990]: I1203 12:53:58.494766 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerStarted","Data":"08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac"} Dec 03 12:53:59 crc kubenswrapper[4990]: I1203 12:53:59.428933 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hvzjl"] Dec 03 12:53:59 crc kubenswrapper[4990]: I1203 12:53:59.502545 4990 generic.go:334] "Generic (PLEG): container finished" podID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerID="08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac" exitCode=0 Dec 03 12:53:59 crc kubenswrapper[4990]: I1203 12:53:59.502606 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerDied","Data":"08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac"} Dec 03 12:53:59 crc kubenswrapper[4990]: I1203 12:53:59.502785 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hvzjl" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="registry-server" containerID="cri-o://37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e" gracePeriod=2 Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.406546 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.512615 4990 generic.go:334] "Generic (PLEG): container finished" podID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerID="37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e" exitCode=0 Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.512679 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerDied","Data":"37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e"} Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.512741 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hvzjl" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.512781 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hvzjl" event={"ID":"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8","Type":"ContainerDied","Data":"17dc8b06cee3c6c1a16b8a0676804cf8f8cd754407e34a0fda47071d4b579aa5"} Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.512814 4990 scope.go:117] "RemoveContainer" containerID="37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.529241 4990 scope.go:117] "RemoveContainer" containerID="399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.541986 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-catalog-content\") pod \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.542070 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbxl2\" (UniqueName: \"kubernetes.io/projected/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-kube-api-access-fbxl2\") pod \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.542139 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-utilities\") pod \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\" (UID: \"74872f7d-97bf-4d26-8ef3-2b2984c6b0e8\") " Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.543256 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-utilities" (OuterVolumeSpecName: "utilities") pod "74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" (UID: "74872f7d-97bf-4d26-8ef3-2b2984c6b0e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.556048 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-kube-api-access-fbxl2" (OuterVolumeSpecName: "kube-api-access-fbxl2") pod "74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" (UID: "74872f7d-97bf-4d26-8ef3-2b2984c6b0e8"). InnerVolumeSpecName "kube-api-access-fbxl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.562074 4990 scope.go:117] "RemoveContainer" containerID="d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.583576 4990 scope.go:117] "RemoveContainer" containerID="37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e" Dec 03 12:54:00 crc kubenswrapper[4990]: E1203 12:54:00.584227 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e\": container with ID starting with 37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e not found: ID does not exist" containerID="37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.584259 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e"} err="failed to get container status \"37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e\": rpc error: code = NotFound desc = could not find container \"37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e\": container with ID starting with 37da996028d4d44a60fa9605c896c89a8aba1ddb3fc8ea35f269dccebfcef49e not found: ID does not exist" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.584281 4990 scope.go:117] "RemoveContainer" containerID="399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f" Dec 03 12:54:00 crc kubenswrapper[4990]: E1203 12:54:00.584798 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f\": container with ID starting with 399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f not found: ID does not exist" containerID="399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.584824 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f"} err="failed to get container status \"399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f\": rpc error: code = NotFound desc = could not find container \"399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f\": container with ID starting with 399704028e974853ab3720caf68f8189632b69fda1bb66a9b326dc0cfe5a686f not found: ID does not exist" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.584837 4990 scope.go:117] "RemoveContainer" containerID="d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c" Dec 03 12:54:00 crc kubenswrapper[4990]: E1203 12:54:00.585157 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c\": container with ID starting with d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c not found: ID does not exist" containerID="d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.585181 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c"} err="failed to get container status \"d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c\": rpc error: code = NotFound desc = could not find container \"d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c\": container with ID starting with d62f09bab0fa9f60d78e977b29235b36ced46f65d0a3c03633bca3f6db74998c not found: ID does not exist" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.615222 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" (UID: "74872f7d-97bf-4d26-8ef3-2b2984c6b0e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.644405 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbxl2\" (UniqueName: \"kubernetes.io/projected/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-kube-api-access-fbxl2\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.644442 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.644474 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.868221 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hvzjl"] Dec 03 12:54:00 crc kubenswrapper[4990]: I1203 12:54:00.878432 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hvzjl"] Dec 03 12:54:01 crc kubenswrapper[4990]: I1203 12:54:01.521203 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerStarted","Data":"4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896"} Dec 03 12:54:01 crc kubenswrapper[4990]: I1203 12:54:01.538692 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f7w54" podStartSLOduration=3.875729319 podStartE2EDuration="6.538676274s" podCreationTimestamp="2025-12-03 12:53:55 +0000 UTC" firstStartedPulling="2025-12-03 12:53:57.489662642 +0000 UTC m=+985.631573871" lastFinishedPulling="2025-12-03 12:54:00.152609597 +0000 UTC m=+988.294520826" observedRunningTime="2025-12-03 12:54:01.536660651 +0000 UTC m=+989.678571910" watchObservedRunningTime="2025-12-03 12:54:01.538676274 +0000 UTC m=+989.680587503" Dec 03 12:54:02 crc kubenswrapper[4990]: I1203 12:54:02.271547 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" path="/var/lib/kubelet/pods/74872f7d-97bf-4d26-8ef3-2b2984c6b0e8/volumes" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.087161 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-584cf47849-n84tl" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.897900 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-p4dt2"] Dec 03 12:54:04 crc kubenswrapper[4990]: E1203 12:54:04.898192 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="extract-utilities" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.898209 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="extract-utilities" Dec 03 12:54:04 crc kubenswrapper[4990]: E1203 12:54:04.898229 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="registry-server" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.898237 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="registry-server" Dec 03 12:54:04 crc kubenswrapper[4990]: E1203 12:54:04.898246 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="extract-content" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.898253 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="extract-content" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.898409 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="74872f7d-97bf-4d26-8ef3-2b2984c6b0e8" containerName="registry-server" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.900433 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.902312 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-t666x" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.902645 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.903090 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.911936 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6"] Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.912958 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.915837 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 12:54:04 crc kubenswrapper[4990]: I1203 12:54:04.932502 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6"] Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002612 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-reloader\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002661 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002694 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-sockets\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002719 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-conf\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002734 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics-certs\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002771 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-startup\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.002788 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xm9c\" (UniqueName: \"kubernetes.io/projected/f6a369fa-06fe-4bb3-a796-c54b8442582e-kube-api-access-7xm9c\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.030344 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-t6cl8"] Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.031821 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.038429 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-p92rn" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.038661 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.038790 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.038903 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.044201 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-hhxt8"] Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.046379 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.049687 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.060905 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-hhxt8"] Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.103652 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-conf\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.103714 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics-certs\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.103755 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d64e42b3-c578-44d9-818b-3b0aa7f4b3e4-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-xnhk6\" (UID: \"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: E1203 12:54:05.103882 4990 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.103899 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-startup\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: E1203 12:54:05.103957 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics-certs podName:f6a369fa-06fe-4bb3-a796-c54b8442582e nodeName:}" failed. No retries permitted until 2025-12-03 12:54:05.60393489 +0000 UTC m=+993.745846189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics-certs") pod "frr-k8s-p4dt2" (UID: "f6a369fa-06fe-4bb3-a796-c54b8442582e") : secret "frr-k8s-certs-secret" not found Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.103990 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xm9c\" (UniqueName: \"kubernetes.io/projected/f6a369fa-06fe-4bb3-a796-c54b8442582e-kube-api-access-7xm9c\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104108 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-reloader\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104525 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-conf\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104578 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104721 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvl49\" (UniqueName: \"kubernetes.io/projected/d64e42b3-c578-44d9-818b-3b0aa7f4b3e4-kube-api-access-hvl49\") pod \"frr-k8s-webhook-server-7fcb986d4-xnhk6\" (UID: \"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104768 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-sockets\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104848 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-reloader\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.104894 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-startup\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.105068 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-frr-sockets\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.105167 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.124758 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xm9c\" (UniqueName: \"kubernetes.io/projected/f6a369fa-06fe-4bb3-a796-c54b8442582e-kube-api-access-7xm9c\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205144 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d94113a-0245-4057-97e6-8464ef2a830d-metrics-certs\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205525 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvl49\" (UniqueName: \"kubernetes.io/projected/d64e42b3-c578-44d9-818b-3b0aa7f4b3e4-kube-api-access-hvl49\") pod \"frr-k8s-webhook-server-7fcb986d4-xnhk6\" (UID: \"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205549 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-metrics-certs\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205568 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knvhf\" (UniqueName: \"kubernetes.io/projected/9d94113a-0245-4057-97e6-8464ef2a830d-kube-api-access-knvhf\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205586 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9d94113a-0245-4057-97e6-8464ef2a830d-cert\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205611 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-metallb-excludel2\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205637 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205665 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d64e42b3-c578-44d9-818b-3b0aa7f4b3e4-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-xnhk6\" (UID: \"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.205707 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trdmc\" (UniqueName: \"kubernetes.io/projected/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-kube-api-access-trdmc\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.212616 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d64e42b3-c578-44d9-818b-3b0aa7f4b3e4-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-xnhk6\" (UID: \"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.221695 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvl49\" (UniqueName: \"kubernetes.io/projected/d64e42b3-c578-44d9-818b-3b0aa7f4b3e4-kube-api-access-hvl49\") pod \"frr-k8s-webhook-server-7fcb986d4-xnhk6\" (UID: \"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.229160 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306290 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trdmc\" (UniqueName: \"kubernetes.io/projected/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-kube-api-access-trdmc\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306348 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d94113a-0245-4057-97e6-8464ef2a830d-metrics-certs\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306383 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-metrics-certs\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306412 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knvhf\" (UniqueName: \"kubernetes.io/projected/9d94113a-0245-4057-97e6-8464ef2a830d-kube-api-access-knvhf\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306808 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9d94113a-0245-4057-97e6-8464ef2a830d-cert\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306864 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-metallb-excludel2\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.306889 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: E1203 12:54:05.307063 4990 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 12:54:05 crc kubenswrapper[4990]: E1203 12:54:05.307122 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist podName:10554cf0-1dc9-4e97-973f-1f2d47e3d1f8 nodeName:}" failed. No retries permitted until 2025-12-03 12:54:05.807106117 +0000 UTC m=+993.949017346 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist") pod "speaker-t6cl8" (UID: "10554cf0-1dc9-4e97-973f-1f2d47e3d1f8") : secret "metallb-memberlist" not found Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.307701 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-metallb-excludel2\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.308677 4990 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.311567 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-metrics-certs\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.327819 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9d94113a-0245-4057-97e6-8464ef2a830d-cert\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.327907 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trdmc\" (UniqueName: \"kubernetes.io/projected/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-kube-api-access-trdmc\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.334375 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knvhf\" (UniqueName: \"kubernetes.io/projected/9d94113a-0245-4057-97e6-8464ef2a830d-kube-api-access-knvhf\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.336147 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d94113a-0245-4057-97e6-8464ef2a830d-metrics-certs\") pod \"controller-f8648f98b-hhxt8\" (UID: \"9d94113a-0245-4057-97e6-8464ef2a830d\") " pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.367107 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.572019 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-hhxt8"] Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.614337 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics-certs\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.624752 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6a369fa-06fe-4bb3-a796-c54b8442582e-metrics-certs\") pod \"frr-k8s-p4dt2\" (UID: \"f6a369fa-06fe-4bb3-a796-c54b8442582e\") " pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.660709 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6"] Dec 03 12:54:05 crc kubenswrapper[4990]: W1203 12:54:05.673644 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd64e42b3_c578_44d9_818b_3b0aa7f4b3e4.slice/crio-5d3b68ccf182ecca1955f5e43e503344541f424b7cf31f3af5258e8c6161153f WatchSource:0}: Error finding container 5d3b68ccf182ecca1955f5e43e503344541f424b7cf31f3af5258e8c6161153f: Status 404 returned error can't find the container with id 5d3b68ccf182ecca1955f5e43e503344541f424b7cf31f3af5258e8c6161153f Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.817272 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:05 crc kubenswrapper[4990]: E1203 12:54:05.817534 4990 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 12:54:05 crc kubenswrapper[4990]: E1203 12:54:05.817927 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist podName:10554cf0-1dc9-4e97-973f-1f2d47e3d1f8 nodeName:}" failed. No retries permitted until 2025-12-03 12:54:06.817907766 +0000 UTC m=+994.959818995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist") pod "speaker-t6cl8" (UID: "10554cf0-1dc9-4e97-973f-1f2d47e3d1f8") : secret "metallb-memberlist" not found Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.820479 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.961005 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:54:05 crc kubenswrapper[4990]: I1203 12:54:05.961044 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.003046 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.555776 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hhxt8" event={"ID":"9d94113a-0245-4057-97e6-8464ef2a830d","Type":"ContainerStarted","Data":"0b505ea49744679ca5f92c792d7d325dfc252aa6e679aff48287cb52b0915da8"} Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.556875 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hhxt8" event={"ID":"9d94113a-0245-4057-97e6-8464ef2a830d","Type":"ContainerStarted","Data":"d7f7c1e0e3e8e8e91239f07428e104678861da83ac8ebdbcbc4a18e3bbfb9b72"} Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.556971 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hhxt8" event={"ID":"9d94113a-0245-4057-97e6-8464ef2a830d","Type":"ContainerStarted","Data":"b46599a92775927a088b3a0b7ad857a7d02ef9be270be13c5be7099c9a58f962"} Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.557063 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.557148 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" event={"ID":"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4","Type":"ContainerStarted","Data":"5d3b68ccf182ecca1955f5e43e503344541f424b7cf31f3af5258e8c6161153f"} Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.558328 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"cd1cbff6f04b54be5e2587de0d91a35380ac31385784cf9b2a95fd70b66dfafa"} Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.580488 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-hhxt8" podStartSLOduration=1.580466483 podStartE2EDuration="1.580466483s" podCreationTimestamp="2025-12-03 12:54:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:54:06.578967673 +0000 UTC m=+994.720878922" watchObservedRunningTime="2025-12-03 12:54:06.580466483 +0000 UTC m=+994.722377712" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.626134 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.689618 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f7w54"] Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.829939 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.838299 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10554cf0-1dc9-4e97-973f-1f2d47e3d1f8-memberlist\") pod \"speaker-t6cl8\" (UID: \"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8\") " pod="metallb-system/speaker-t6cl8" Dec 03 12:54:06 crc kubenswrapper[4990]: I1203 12:54:06.847247 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-t6cl8" Dec 03 12:54:07 crc kubenswrapper[4990]: I1203 12:54:07.568872 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-t6cl8" event={"ID":"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8","Type":"ContainerStarted","Data":"559167ddfbaf127bbe098a9841f49d2e848033a42fd64faa4fa699a204760d47"} Dec 03 12:54:07 crc kubenswrapper[4990]: I1203 12:54:07.568930 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-t6cl8" event={"ID":"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8","Type":"ContainerStarted","Data":"7209bdfeed9be2ff8a72154c5f492e41ae3b915d87bf551b49645c24991bfb8e"} Dec 03 12:54:07 crc kubenswrapper[4990]: I1203 12:54:07.568944 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-t6cl8" event={"ID":"10554cf0-1dc9-4e97-973f-1f2d47e3d1f8","Type":"ContainerStarted","Data":"b295a5de03533ae9401fe8730d2d961ea378bfd0a85178054f5b404a7379a7d7"} Dec 03 12:54:07 crc kubenswrapper[4990]: I1203 12:54:07.589067 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-t6cl8" podStartSLOduration=2.589048466 podStartE2EDuration="2.589048466s" podCreationTimestamp="2025-12-03 12:54:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:54:07.586064267 +0000 UTC m=+995.727975506" watchObservedRunningTime="2025-12-03 12:54:07.589048466 +0000 UTC m=+995.730959695" Dec 03 12:54:08 crc kubenswrapper[4990]: I1203 12:54:08.574417 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f7w54" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="registry-server" containerID="cri-o://4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896" gracePeriod=2 Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.023317 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.168888 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-utilities\") pod \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.169117 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-catalog-content\") pod \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.169181 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5nsq\" (UniqueName: \"kubernetes.io/projected/c8dc90fc-14d0-4d60-8cff-6b58658252e0-kube-api-access-n5nsq\") pod \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\" (UID: \"c8dc90fc-14d0-4d60-8cff-6b58658252e0\") " Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.170100 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-utilities" (OuterVolumeSpecName: "utilities") pod "c8dc90fc-14d0-4d60-8cff-6b58658252e0" (UID: "c8dc90fc-14d0-4d60-8cff-6b58658252e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.173290 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.179922 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dc90fc-14d0-4d60-8cff-6b58658252e0-kube-api-access-n5nsq" (OuterVolumeSpecName: "kube-api-access-n5nsq") pod "c8dc90fc-14d0-4d60-8cff-6b58658252e0" (UID: "c8dc90fc-14d0-4d60-8cff-6b58658252e0"). InnerVolumeSpecName "kube-api-access-n5nsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.234091 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8dc90fc-14d0-4d60-8cff-6b58658252e0" (UID: "c8dc90fc-14d0-4d60-8cff-6b58658252e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.274365 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8dc90fc-14d0-4d60-8cff-6b58658252e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.274405 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5nsq\" (UniqueName: \"kubernetes.io/projected/c8dc90fc-14d0-4d60-8cff-6b58658252e0-kube-api-access-n5nsq\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.585599 4990 generic.go:334] "Generic (PLEG): container finished" podID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerID="4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896" exitCode=0 Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.585676 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerDied","Data":"4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896"} Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.585706 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7w54" event={"ID":"c8dc90fc-14d0-4d60-8cff-6b58658252e0","Type":"ContainerDied","Data":"4d6063bd8f02aacf40330207c62f019c43e13fd6271711d36959d307468b09c7"} Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.585744 4990 scope.go:117] "RemoveContainer" containerID="4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.585931 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7w54" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.632141 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f7w54"] Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.636432 4990 scope.go:117] "RemoveContainer" containerID="08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.639738 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f7w54"] Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.659997 4990 scope.go:117] "RemoveContainer" containerID="e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.696036 4990 scope.go:117] "RemoveContainer" containerID="4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896" Dec 03 12:54:09 crc kubenswrapper[4990]: E1203 12:54:09.697154 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896\": container with ID starting with 4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896 not found: ID does not exist" containerID="4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.697204 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896"} err="failed to get container status \"4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896\": rpc error: code = NotFound desc = could not find container \"4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896\": container with ID starting with 4a9666559a121fe45c3348712f4b82e2cd40961c58f27fead89400b6c003f896 not found: ID does not exist" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.697243 4990 scope.go:117] "RemoveContainer" containerID="08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac" Dec 03 12:54:09 crc kubenswrapper[4990]: E1203 12:54:09.698082 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac\": container with ID starting with 08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac not found: ID does not exist" containerID="08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.698113 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac"} err="failed to get container status \"08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac\": rpc error: code = NotFound desc = could not find container \"08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac\": container with ID starting with 08453e31a8bf16674655f280219219a223f920b3bf03976b99fe82ebf49a65ac not found: ID does not exist" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.698137 4990 scope.go:117] "RemoveContainer" containerID="e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632" Dec 03 12:54:09 crc kubenswrapper[4990]: E1203 12:54:09.698968 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632\": container with ID starting with e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632 not found: ID does not exist" containerID="e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632" Dec 03 12:54:09 crc kubenswrapper[4990]: I1203 12:54:09.699209 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632"} err="failed to get container status \"e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632\": rpc error: code = NotFound desc = could not find container \"e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632\": container with ID starting with e8de4272079e39c98960bc1cc491f167c6803cd4af0168d1a52b70518ef84632 not found: ID does not exist" Dec 03 12:54:10 crc kubenswrapper[4990]: I1203 12:54:10.271691 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" path="/var/lib/kubelet/pods/c8dc90fc-14d0-4d60-8cff-6b58658252e0/volumes" Dec 03 12:54:15 crc kubenswrapper[4990]: I1203 12:54:15.642589 4990 generic.go:334] "Generic (PLEG): container finished" podID="f6a369fa-06fe-4bb3-a796-c54b8442582e" containerID="483a523cffbf6e1dc77ae38ebfdb45a3bbd1ab32146ee637b43d113153b7bf49" exitCode=0 Dec 03 12:54:15 crc kubenswrapper[4990]: I1203 12:54:15.643563 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerDied","Data":"483a523cffbf6e1dc77ae38ebfdb45a3bbd1ab32146ee637b43d113153b7bf49"} Dec 03 12:54:15 crc kubenswrapper[4990]: I1203 12:54:15.647388 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" event={"ID":"d64e42b3-c578-44d9-818b-3b0aa7f4b3e4","Type":"ContainerStarted","Data":"a0525423618633421b65e014d3ad8cf53e46cfbdfeab4040ff5f722ea75af843"} Dec 03 12:54:15 crc kubenswrapper[4990]: I1203 12:54:15.656816 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:15 crc kubenswrapper[4990]: I1203 12:54:15.711116 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" podStartSLOduration=2.754367287 podStartE2EDuration="11.711092332s" podCreationTimestamp="2025-12-03 12:54:04 +0000 UTC" firstStartedPulling="2025-12-03 12:54:05.675886661 +0000 UTC m=+993.817797890" lastFinishedPulling="2025-12-03 12:54:14.632611706 +0000 UTC m=+1002.774522935" observedRunningTime="2025-12-03 12:54:15.709625903 +0000 UTC m=+1003.851537132" watchObservedRunningTime="2025-12-03 12:54:15.711092332 +0000 UTC m=+1003.853003571" Dec 03 12:54:16 crc kubenswrapper[4990]: I1203 12:54:16.655901 4990 generic.go:334] "Generic (PLEG): container finished" podID="f6a369fa-06fe-4bb3-a796-c54b8442582e" containerID="c21bea1ddbdc18d2502c24fd97af89d5e9a615d301367e01eff9cb08b06d2ffc" exitCode=0 Dec 03 12:54:16 crc kubenswrapper[4990]: I1203 12:54:16.655967 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerDied","Data":"c21bea1ddbdc18d2502c24fd97af89d5e9a615d301367e01eff9cb08b06d2ffc"} Dec 03 12:54:16 crc kubenswrapper[4990]: I1203 12:54:16.847538 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-t6cl8" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.028138 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rvbv9"] Dec 03 12:54:17 crc kubenswrapper[4990]: E1203 12:54:17.028498 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="extract-content" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.028514 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="extract-content" Dec 03 12:54:17 crc kubenswrapper[4990]: E1203 12:54:17.028538 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="extract-utilities" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.028546 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="extract-utilities" Dec 03 12:54:17 crc kubenswrapper[4990]: E1203 12:54:17.028565 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="registry-server" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.028578 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="registry-server" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.028686 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8dc90fc-14d0-4d60-8cff-6b58658252e0" containerName="registry-server" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.029619 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.043756 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvbv9"] Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.112283 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spdn6\" (UniqueName: \"kubernetes.io/projected/23ac53b6-73de-4cdb-8223-da8f04af923c-kube-api-access-spdn6\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.112360 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-catalog-content\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.112394 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-utilities\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.214019 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spdn6\" (UniqueName: \"kubernetes.io/projected/23ac53b6-73de-4cdb-8223-da8f04af923c-kube-api-access-spdn6\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.214122 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-catalog-content\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.214155 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-utilities\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.214897 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-utilities\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.214955 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-catalog-content\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.237759 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spdn6\" (UniqueName: \"kubernetes.io/projected/23ac53b6-73de-4cdb-8223-da8f04af923c-kube-api-access-spdn6\") pod \"redhat-marketplace-rvbv9\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.349956 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.655015 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvbv9"] Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.664867 4990 generic.go:334] "Generic (PLEG): container finished" podID="f6a369fa-06fe-4bb3-a796-c54b8442582e" containerID="584a5eb0d9da460af17c2a0b4bc451e65eb774041ac36fa76cd28f2cf5bc4d63" exitCode=0 Dec 03 12:54:17 crc kubenswrapper[4990]: I1203 12:54:17.665572 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerDied","Data":"584a5eb0d9da460af17c2a0b4bc451e65eb774041ac36fa76cd28f2cf5bc4d63"} Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.677078 4990 generic.go:334] "Generic (PLEG): container finished" podID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerID="7f4680f60c9bb8422e249f901e06b55f521143dcc19daf4dddb6f21db777f599" exitCode=0 Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.677195 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvbv9" event={"ID":"23ac53b6-73de-4cdb-8223-da8f04af923c","Type":"ContainerDied","Data":"7f4680f60c9bb8422e249f901e06b55f521143dcc19daf4dddb6f21db777f599"} Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.678220 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvbv9" event={"ID":"23ac53b6-73de-4cdb-8223-da8f04af923c","Type":"ContainerStarted","Data":"b028234a3b05b6e253c59b834ff3959dd3c364a6aeb030bf4a699659b04deaa2"} Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.689913 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"c282db5c83f4c5da37d858bcc198e8fe2084aa7ce460ce5c846d0d61c9b42864"} Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.689972 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"fd0b0dbd62b77a768432b7d3fc28fd3e44492a832b35ad0aa7a4c7f812059292"} Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.689993 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"186f4bd53ec8e7727dbc542da3eb27950d34722a4433321b9a68531c6b27bd23"} Dec 03 12:54:18 crc kubenswrapper[4990]: I1203 12:54:18.690007 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"2a33b75a23dcce59fdffcf7d76aaca8b5ee6e1379741198d06f83d40358ece71"} Dec 03 12:54:19 crc kubenswrapper[4990]: I1203 12:54:19.705268 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"b622a30f2f8378e4abcca2a5aaec8dbcd5648a1f1922a49304ee890901e83ede"} Dec 03 12:54:19 crc kubenswrapper[4990]: I1203 12:54:19.705743 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p4dt2" event={"ID":"f6a369fa-06fe-4bb3-a796-c54b8442582e","Type":"ContainerStarted","Data":"20694c91df2f00c39120c36fb61f1fcd950f43ea2d1f318035c83670880069a9"} Dec 03 12:54:19 crc kubenswrapper[4990]: I1203 12:54:19.705829 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:19 crc kubenswrapper[4990]: I1203 12:54:19.738995 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-p4dt2" podStartSLOduration=7.230987623 podStartE2EDuration="15.738963276s" podCreationTimestamp="2025-12-03 12:54:04 +0000 UTC" firstStartedPulling="2025-12-03 12:54:06.142998987 +0000 UTC m=+994.284910226" lastFinishedPulling="2025-12-03 12:54:14.65097465 +0000 UTC m=+1002.792885879" observedRunningTime="2025-12-03 12:54:19.737477237 +0000 UTC m=+1007.879388466" watchObservedRunningTime="2025-12-03 12:54:19.738963276 +0000 UTC m=+1007.880874515" Dec 03 12:54:20 crc kubenswrapper[4990]: I1203 12:54:20.716727 4990 generic.go:334] "Generic (PLEG): container finished" podID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerID="a06b81891d9a694411581890ab8a351ba9b12f92cd4e6bcd90af0b74ba89bbeb" exitCode=0 Dec 03 12:54:20 crc kubenswrapper[4990]: I1203 12:54:20.716787 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvbv9" event={"ID":"23ac53b6-73de-4cdb-8223-da8f04af923c","Type":"ContainerDied","Data":"a06b81891d9a694411581890ab8a351ba9b12f92cd4e6bcd90af0b74ba89bbeb"} Dec 03 12:54:20 crc kubenswrapper[4990]: I1203 12:54:20.821406 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:20 crc kubenswrapper[4990]: I1203 12:54:20.936842 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:21 crc kubenswrapper[4990]: I1203 12:54:21.726979 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvbv9" event={"ID":"23ac53b6-73de-4cdb-8223-da8f04af923c","Type":"ContainerStarted","Data":"04a8b6e4fe2faa7a609fe2813c2415fc71d1c8fe1b559e9614b76bcbc7c84675"} Dec 03 12:54:21 crc kubenswrapper[4990]: I1203 12:54:21.748003 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rvbv9" podStartSLOduration=2.3035639359999998 podStartE2EDuration="4.747978248s" podCreationTimestamp="2025-12-03 12:54:17 +0000 UTC" firstStartedPulling="2025-12-03 12:54:18.68098694 +0000 UTC m=+1006.822898159" lastFinishedPulling="2025-12-03 12:54:21.125401242 +0000 UTC m=+1009.267312471" observedRunningTime="2025-12-03 12:54:21.746766446 +0000 UTC m=+1009.888677685" watchObservedRunningTime="2025-12-03 12:54:21.747978248 +0000 UTC m=+1009.889889477" Dec 03 12:54:25 crc kubenswrapper[4990]: I1203 12:54:25.239772 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-xnhk6" Dec 03 12:54:25 crc kubenswrapper[4990]: I1203 12:54:25.371877 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-hhxt8" Dec 03 12:54:26 crc kubenswrapper[4990]: I1203 12:54:26.850823 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-t6cl8" Dec 03 12:54:27 crc kubenswrapper[4990]: I1203 12:54:27.350844 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:27 crc kubenswrapper[4990]: I1203 12:54:27.351119 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:27 crc kubenswrapper[4990]: I1203 12:54:27.394417 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:27 crc kubenswrapper[4990]: I1203 12:54:27.813759 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:27 crc kubenswrapper[4990]: I1203 12:54:27.874232 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvbv9"] Dec 03 12:54:29 crc kubenswrapper[4990]: I1203 12:54:29.785537 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rvbv9" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="registry-server" containerID="cri-o://04a8b6e4fe2faa7a609fe2813c2415fc71d1c8fe1b559e9614b76bcbc7c84675" gracePeriod=2 Dec 03 12:54:31 crc kubenswrapper[4990]: I1203 12:54:31.808896 4990 generic.go:334] "Generic (PLEG): container finished" podID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerID="04a8b6e4fe2faa7a609fe2813c2415fc71d1c8fe1b559e9614b76bcbc7c84675" exitCode=0 Dec 03 12:54:31 crc kubenswrapper[4990]: I1203 12:54:31.808951 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvbv9" event={"ID":"23ac53b6-73de-4cdb-8223-da8f04af923c","Type":"ContainerDied","Data":"04a8b6e4fe2faa7a609fe2813c2415fc71d1c8fe1b559e9614b76bcbc7c84675"} Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.448614 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xgxsw"] Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.449491 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.451874 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.452234 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.452705 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7bsbq" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.463844 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xgxsw"] Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.482746 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrhd2\" (UniqueName: \"kubernetes.io/projected/c1b91baf-8945-490b-987c-531d89259d3b-kube-api-access-vrhd2\") pod \"openstack-operator-index-xgxsw\" (UID: \"c1b91baf-8945-490b-987c-531d89259d3b\") " pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.584692 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrhd2\" (UniqueName: \"kubernetes.io/projected/c1b91baf-8945-490b-987c-531d89259d3b-kube-api-access-vrhd2\") pod \"openstack-operator-index-xgxsw\" (UID: \"c1b91baf-8945-490b-987c-531d89259d3b\") " pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.609241 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrhd2\" (UniqueName: \"kubernetes.io/projected/c1b91baf-8945-490b-987c-531d89259d3b-kube-api-access-vrhd2\") pod \"openstack-operator-index-xgxsw\" (UID: \"c1b91baf-8945-490b-987c-531d89259d3b\") " pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.774470 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.825506 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rvbv9" event={"ID":"23ac53b6-73de-4cdb-8223-da8f04af923c","Type":"ContainerDied","Data":"b028234a3b05b6e253c59b834ff3959dd3c364a6aeb030bf4a699659b04deaa2"} Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.825577 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b028234a3b05b6e253c59b834ff3959dd3c364a6aeb030bf4a699659b04deaa2" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.880755 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.990035 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-catalog-content\") pod \"23ac53b6-73de-4cdb-8223-da8f04af923c\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.990133 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-utilities\") pod \"23ac53b6-73de-4cdb-8223-da8f04af923c\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.990302 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spdn6\" (UniqueName: \"kubernetes.io/projected/23ac53b6-73de-4cdb-8223-da8f04af923c-kube-api-access-spdn6\") pod \"23ac53b6-73de-4cdb-8223-da8f04af923c\" (UID: \"23ac53b6-73de-4cdb-8223-da8f04af923c\") " Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.992071 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-utilities" (OuterVolumeSpecName: "utilities") pod "23ac53b6-73de-4cdb-8223-da8f04af923c" (UID: "23ac53b6-73de-4cdb-8223-da8f04af923c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:32 crc kubenswrapper[4990]: I1203 12:54:32.997308 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ac53b6-73de-4cdb-8223-da8f04af923c-kube-api-access-spdn6" (OuterVolumeSpecName: "kube-api-access-spdn6") pod "23ac53b6-73de-4cdb-8223-da8f04af923c" (UID: "23ac53b6-73de-4cdb-8223-da8f04af923c"). InnerVolumeSpecName "kube-api-access-spdn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.008734 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23ac53b6-73de-4cdb-8223-da8f04af923c" (UID: "23ac53b6-73de-4cdb-8223-da8f04af923c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.092237 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.092293 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ac53b6-73de-4cdb-8223-da8f04af923c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.092307 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spdn6\" (UniqueName: \"kubernetes.io/projected/23ac53b6-73de-4cdb-8223-da8f04af923c-kube-api-access-spdn6\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.214813 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xgxsw"] Dec 03 12:54:33 crc kubenswrapper[4990]: W1203 12:54:33.226869 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1b91baf_8945_490b_987c_531d89259d3b.slice/crio-c9aebc92d5bdb51bc8b3f267d73fa5618b508b15ba252e3d6cb5d382a107e92d WatchSource:0}: Error finding container c9aebc92d5bdb51bc8b3f267d73fa5618b508b15ba252e3d6cb5d382a107e92d: Status 404 returned error can't find the container with id c9aebc92d5bdb51bc8b3f267d73fa5618b508b15ba252e3d6cb5d382a107e92d Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.834058 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xgxsw" event={"ID":"c1b91baf-8945-490b-987c-531d89259d3b","Type":"ContainerStarted","Data":"c9aebc92d5bdb51bc8b3f267d73fa5618b508b15ba252e3d6cb5d382a107e92d"} Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.834084 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rvbv9" Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.872635 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvbv9"] Dec 03 12:54:33 crc kubenswrapper[4990]: I1203 12:54:33.880345 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rvbv9"] Dec 03 12:54:34 crc kubenswrapper[4990]: I1203 12:54:34.277975 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" path="/var/lib/kubelet/pods/23ac53b6-73de-4cdb-8223-da8f04af923c/volumes" Dec 03 12:54:35 crc kubenswrapper[4990]: I1203 12:54:35.824214 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-p4dt2" Dec 03 12:54:38 crc kubenswrapper[4990]: I1203 12:54:38.878181 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xgxsw" event={"ID":"c1b91baf-8945-490b-987c-531d89259d3b","Type":"ContainerStarted","Data":"5f9820c2f21020ddc7c8423c4da247b990ad1b3feb851d1eceef9d12cb318ffa"} Dec 03 12:54:38 crc kubenswrapper[4990]: I1203 12:54:38.902053 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xgxsw" podStartSLOduration=2.537315788 podStartE2EDuration="6.902025254s" podCreationTimestamp="2025-12-03 12:54:32 +0000 UTC" firstStartedPulling="2025-12-03 12:54:33.230384908 +0000 UTC m=+1021.372296147" lastFinishedPulling="2025-12-03 12:54:37.595094384 +0000 UTC m=+1025.737005613" observedRunningTime="2025-12-03 12:54:38.897206208 +0000 UTC m=+1027.039117457" watchObservedRunningTime="2025-12-03 12:54:38.902025254 +0000 UTC m=+1027.043936483" Dec 03 12:54:42 crc kubenswrapper[4990]: I1203 12:54:42.775721 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:42 crc kubenswrapper[4990]: I1203 12:54:42.776148 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:42 crc kubenswrapper[4990]: I1203 12:54:42.816763 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:43 crc kubenswrapper[4990]: I1203 12:54:43.939552 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xgxsw" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.094174 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm"] Dec 03 12:54:51 crc kubenswrapper[4990]: E1203 12:54:51.095194 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="extract-content" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.095212 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="extract-content" Dec 03 12:54:51 crc kubenswrapper[4990]: E1203 12:54:51.095219 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="extract-utilities" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.095225 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="extract-utilities" Dec 03 12:54:51 crc kubenswrapper[4990]: E1203 12:54:51.095238 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="registry-server" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.095245 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="registry-server" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.095402 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ac53b6-73de-4cdb-8223-da8f04af923c" containerName="registry-server" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.096367 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.099035 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-55phl" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.108160 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm"] Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.219093 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-bundle\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.219373 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7xdj\" (UniqueName: \"kubernetes.io/projected/50576f83-b05b-418b-a57d-0a322f2c489f-kube-api-access-j7xdj\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.219544 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-util\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.321754 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-bundle\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.321894 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7xdj\" (UniqueName: \"kubernetes.io/projected/50576f83-b05b-418b-a57d-0a322f2c489f-kube-api-access-j7xdj\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.321962 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-util\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.322807 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-bundle\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.322901 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-util\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.344837 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7xdj\" (UniqueName: \"kubernetes.io/projected/50576f83-b05b-418b-a57d-0a322f2c489f-kube-api-access-j7xdj\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.428242 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.657444 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm"] Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.969250 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" event={"ID":"50576f83-b05b-418b-a57d-0a322f2c489f","Type":"ContainerStarted","Data":"bb8e1781d622bc69e35bc331f855ee69ffaa205d43a061c58d0e62b87737db4e"} Dec 03 12:54:51 crc kubenswrapper[4990]: I1203 12:54:51.969324 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" event={"ID":"50576f83-b05b-418b-a57d-0a322f2c489f","Type":"ContainerStarted","Data":"9115e916cd9a75660774987f11a24b8f61d67d774b0c1b2062e33218c86647d3"} Dec 03 12:54:52 crc kubenswrapper[4990]: I1203 12:54:52.977476 4990 generic.go:334] "Generic (PLEG): container finished" podID="50576f83-b05b-418b-a57d-0a322f2c489f" containerID="bb8e1781d622bc69e35bc331f855ee69ffaa205d43a061c58d0e62b87737db4e" exitCode=0 Dec 03 12:54:52 crc kubenswrapper[4990]: I1203 12:54:52.977564 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" event={"ID":"50576f83-b05b-418b-a57d-0a322f2c489f","Type":"ContainerDied","Data":"bb8e1781d622bc69e35bc331f855ee69ffaa205d43a061c58d0e62b87737db4e"} Dec 03 12:54:54 crc kubenswrapper[4990]: I1203 12:54:54.998983 4990 generic.go:334] "Generic (PLEG): container finished" podID="50576f83-b05b-418b-a57d-0a322f2c489f" containerID="169fdae462990bc4cc121ccb44d43309add05abd6b9ef147299d05eb376d1055" exitCode=0 Dec 03 12:54:54 crc kubenswrapper[4990]: I1203 12:54:54.999081 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" event={"ID":"50576f83-b05b-418b-a57d-0a322f2c489f","Type":"ContainerDied","Data":"169fdae462990bc4cc121ccb44d43309add05abd6b9ef147299d05eb376d1055"} Dec 03 12:54:56 crc kubenswrapper[4990]: I1203 12:54:56.010534 4990 generic.go:334] "Generic (PLEG): container finished" podID="50576f83-b05b-418b-a57d-0a322f2c489f" containerID="a6875131450b08b93e8ae982de1feec69d7b3b036610207c7d1ae1c7500aba79" exitCode=0 Dec 03 12:54:56 crc kubenswrapper[4990]: I1203 12:54:56.010622 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" event={"ID":"50576f83-b05b-418b-a57d-0a322f2c489f","Type":"ContainerDied","Data":"a6875131450b08b93e8ae982de1feec69d7b3b036610207c7d1ae1c7500aba79"} Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.310751 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.375228 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-bundle\") pod \"50576f83-b05b-418b-a57d-0a322f2c489f\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.375367 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7xdj\" (UniqueName: \"kubernetes.io/projected/50576f83-b05b-418b-a57d-0a322f2c489f-kube-api-access-j7xdj\") pod \"50576f83-b05b-418b-a57d-0a322f2c489f\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.375606 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-util\") pod \"50576f83-b05b-418b-a57d-0a322f2c489f\" (UID: \"50576f83-b05b-418b-a57d-0a322f2c489f\") " Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.376148 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-bundle" (OuterVolumeSpecName: "bundle") pod "50576f83-b05b-418b-a57d-0a322f2c489f" (UID: "50576f83-b05b-418b-a57d-0a322f2c489f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.376650 4990 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.382108 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50576f83-b05b-418b-a57d-0a322f2c489f-kube-api-access-j7xdj" (OuterVolumeSpecName: "kube-api-access-j7xdj") pod "50576f83-b05b-418b-a57d-0a322f2c489f" (UID: "50576f83-b05b-418b-a57d-0a322f2c489f"). InnerVolumeSpecName "kube-api-access-j7xdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.478510 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7xdj\" (UniqueName: \"kubernetes.io/projected/50576f83-b05b-418b-a57d-0a322f2c489f-kube-api-access-j7xdj\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.614485 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-util" (OuterVolumeSpecName: "util") pod "50576f83-b05b-418b-a57d-0a322f2c489f" (UID: "50576f83-b05b-418b-a57d-0a322f2c489f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:54:57 crc kubenswrapper[4990]: I1203 12:54:57.681519 4990 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/50576f83-b05b-418b-a57d-0a322f2c489f-util\") on node \"crc\" DevicePath \"\"" Dec 03 12:54:58 crc kubenswrapper[4990]: I1203 12:54:58.028483 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" event={"ID":"50576f83-b05b-418b-a57d-0a322f2c489f","Type":"ContainerDied","Data":"9115e916cd9a75660774987f11a24b8f61d67d774b0c1b2062e33218c86647d3"} Dec 03 12:54:58 crc kubenswrapper[4990]: I1203 12:54:58.028553 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9115e916cd9a75660774987f11a24b8f61d67d774b0c1b2062e33218c86647d3" Dec 03 12:54:58 crc kubenswrapper[4990]: I1203 12:54:58.028567 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.447377 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92"] Dec 03 12:55:04 crc kubenswrapper[4990]: E1203 12:55:04.448613 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="util" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.448628 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="util" Dec 03 12:55:04 crc kubenswrapper[4990]: E1203 12:55:04.448642 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="pull" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.448648 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="pull" Dec 03 12:55:04 crc kubenswrapper[4990]: E1203 12:55:04.448666 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="extract" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.448672 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="extract" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.448807 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="50576f83-b05b-418b-a57d-0a322f2c489f" containerName="extract" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.449346 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.459095 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-h87hx" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.488655 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grn9m\" (UniqueName: \"kubernetes.io/projected/49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d-kube-api-access-grn9m\") pod \"openstack-operator-controller-operator-85bd6bf7bc-zdj92\" (UID: \"49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d\") " pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.516151 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92"] Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.590934 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grn9m\" (UniqueName: \"kubernetes.io/projected/49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d-kube-api-access-grn9m\") pod \"openstack-operator-controller-operator-85bd6bf7bc-zdj92\" (UID: \"49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d\") " pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.619938 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grn9m\" (UniqueName: \"kubernetes.io/projected/49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d-kube-api-access-grn9m\") pod \"openstack-operator-controller-operator-85bd6bf7bc-zdj92\" (UID: \"49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d\") " pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:04 crc kubenswrapper[4990]: I1203 12:55:04.771477 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:05 crc kubenswrapper[4990]: I1203 12:55:05.044228 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92"] Dec 03 12:55:05 crc kubenswrapper[4990]: I1203 12:55:05.091158 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" event={"ID":"49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d","Type":"ContainerStarted","Data":"e676a909fc06d28d68df60da0fa28171faa828e4f470c48500b723e9915dc7db"} Dec 03 12:55:12 crc kubenswrapper[4990]: I1203 12:55:12.149023 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" event={"ID":"49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d","Type":"ContainerStarted","Data":"170a583085c351f7e58d6af1f514cae5216ffeaa0a6d3fdaf8f0a94fbc387245"} Dec 03 12:55:12 crc kubenswrapper[4990]: I1203 12:55:12.150040 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:12 crc kubenswrapper[4990]: I1203 12:55:12.187397 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" podStartSLOduration=1.3345662489999999 podStartE2EDuration="8.187361289s" podCreationTimestamp="2025-12-03 12:55:04 +0000 UTC" firstStartedPulling="2025-12-03 12:55:05.059515177 +0000 UTC m=+1053.201426406" lastFinishedPulling="2025-12-03 12:55:11.912310217 +0000 UTC m=+1060.054221446" observedRunningTime="2025-12-03 12:55:12.179885932 +0000 UTC m=+1060.321797151" watchObservedRunningTime="2025-12-03 12:55:12.187361289 +0000 UTC m=+1060.329272518" Dec 03 12:55:24 crc kubenswrapper[4990]: I1203 12:55:24.776650 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-85bd6bf7bc-zdj92" Dec 03 12:55:33 crc kubenswrapper[4990]: I1203 12:55:33.286134 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:55:33 crc kubenswrapper[4990]: I1203 12:55:33.286749 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.855403 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89"] Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.886422 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k"] Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.887408 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89"] Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.887608 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.888145 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.893438 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc"] Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.910333 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wn5xz" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.910624 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-s4dgg" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.917102 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.921637 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-jb8hh" Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.946734 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k"] Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.963650 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc"] Dec 03 12:55:53 crc kubenswrapper[4990]: I1203 12:55:53.995895 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.007782 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.015381 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-9k65x" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.056139 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.059776 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qlk5\" (UniqueName: \"kubernetes.io/projected/c7a7a3b2-d356-4dfa-b640-ae39aa2d4840-kube-api-access-9qlk5\") pod \"barbican-operator-controller-manager-7d9dfd778-4sp89\" (UID: \"c7a7a3b2-d356-4dfa-b640-ae39aa2d4840\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.059841 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlpzt\" (UniqueName: \"kubernetes.io/projected/ac715eff-4f7f-44b0-b5ca-c5d283edca36-kube-api-access-hlpzt\") pod \"glance-operator-controller-manager-77987cd8cd-z484w\" (UID: \"ac715eff-4f7f-44b0-b5ca-c5d283edca36\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.059864 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qhh6\" (UniqueName: \"kubernetes.io/projected/93f0e562-5973-4c2a-ab48-aa167ae49ffa-kube-api-access-5qhh6\") pod \"designate-operator-controller-manager-78b4bc895b-hqhfc\" (UID: \"93f0e562-5973-4c2a-ab48-aa167ae49ffa\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.059913 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c99xf\" (UniqueName: \"kubernetes.io/projected/c45d3b36-d188-4bb1-b19f-5521709f572a-kube-api-access-c99xf\") pod \"cinder-operator-controller-manager-859b6ccc6-7n65k\" (UID: \"c45d3b36-d188-4bb1-b19f-5521709f572a\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.071002 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.074440 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-2qcz4" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.085600 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.115252 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.133397 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.134934 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.140649 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-wk752" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.162593 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c99xf\" (UniqueName: \"kubernetes.io/projected/c45d3b36-d188-4bb1-b19f-5521709f572a-kube-api-access-c99xf\") pod \"cinder-operator-controller-manager-859b6ccc6-7n65k\" (UID: \"c45d3b36-d188-4bb1-b19f-5521709f572a\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.162675 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qlk5\" (UniqueName: \"kubernetes.io/projected/c7a7a3b2-d356-4dfa-b640-ae39aa2d4840-kube-api-access-9qlk5\") pod \"barbican-operator-controller-manager-7d9dfd778-4sp89\" (UID: \"c7a7a3b2-d356-4dfa-b640-ae39aa2d4840\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.162729 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlpzt\" (UniqueName: \"kubernetes.io/projected/ac715eff-4f7f-44b0-b5ca-c5d283edca36-kube-api-access-hlpzt\") pod \"glance-operator-controller-manager-77987cd8cd-z484w\" (UID: \"ac715eff-4f7f-44b0-b5ca-c5d283edca36\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.162767 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qhh6\" (UniqueName: \"kubernetes.io/projected/93f0e562-5973-4c2a-ab48-aa167ae49ffa-kube-api-access-5qhh6\") pod \"designate-operator-controller-manager-78b4bc895b-hqhfc\" (UID: \"93f0e562-5973-4c2a-ab48-aa167ae49ffa\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.164359 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.169853 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.171138 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.177195 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-7mzrw" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.178198 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.188748 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.217901 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c99xf\" (UniqueName: \"kubernetes.io/projected/c45d3b36-d188-4bb1-b19f-5521709f572a-kube-api-access-c99xf\") pod \"cinder-operator-controller-manager-859b6ccc6-7n65k\" (UID: \"c45d3b36-d188-4bb1-b19f-5521709f572a\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.221032 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qlk5\" (UniqueName: \"kubernetes.io/projected/c7a7a3b2-d356-4dfa-b640-ae39aa2d4840-kube-api-access-9qlk5\") pod \"barbican-operator-controller-manager-7d9dfd778-4sp89\" (UID: \"c7a7a3b2-d356-4dfa-b640-ae39aa2d4840\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.223607 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlpzt\" (UniqueName: \"kubernetes.io/projected/ac715eff-4f7f-44b0-b5ca-c5d283edca36-kube-api-access-hlpzt\") pod \"glance-operator-controller-manager-77987cd8cd-z484w\" (UID: \"ac715eff-4f7f-44b0-b5ca-c5d283edca36\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.231608 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qhh6\" (UniqueName: \"kubernetes.io/projected/93f0e562-5973-4c2a-ab48-aa167ae49ffa-kube-api-access-5qhh6\") pod \"designate-operator-controller-manager-78b4bc895b-hqhfc\" (UID: \"93f0e562-5973-4c2a-ab48-aa167ae49ffa\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.233594 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.258001 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.260006 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.264483 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxc7\" (UniqueName: \"kubernetes.io/projected/a8fb4bb3-4b67-4ada-8876-389bff0e9836-kube-api-access-cnxc7\") pod \"heat-operator-controller-manager-5f64f6f8bb-xt2lq\" (UID: \"a8fb4bb3-4b67-4ada-8876-389bff0e9836\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.264556 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lts6p\" (UniqueName: \"kubernetes.io/projected/4881a1ed-a968-4f20-a8e6-94a91ec5eceb-kube-api-access-lts6p\") pod \"horizon-operator-controller-manager-68c6d99b8f-8zwv5\" (UID: \"4881a1ed-a968-4f20-a8e6-94a91ec5eceb\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.266674 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-wkx66" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.286793 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.323004 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.323934 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.325268 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.325301 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.325753 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.327538 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.327678 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.341286 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.342554 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.349136 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-44t5p" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.367353 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6h2s\" (UniqueName: \"kubernetes.io/projected/0e880282-b396-4fa9-a3ac-89d81315ecd8-kube-api-access-h6h2s\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.367432 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxc7\" (UniqueName: \"kubernetes.io/projected/a8fb4bb3-4b67-4ada-8876-389bff0e9836-kube-api-access-cnxc7\") pod \"heat-operator-controller-manager-5f64f6f8bb-xt2lq\" (UID: \"a8fb4bb3-4b67-4ada-8876-389bff0e9836\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.367479 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bct9w\" (UniqueName: \"kubernetes.io/projected/96da719f-5659-45eb-ac48-37165c936527-kube-api-access-bct9w\") pod \"ironic-operator-controller-manager-6c548fd776-79xxb\" (UID: \"96da719f-5659-45eb-ac48-37165c936527\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.367507 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lts6p\" (UniqueName: \"kubernetes.io/projected/4881a1ed-a968-4f20-a8e6-94a91ec5eceb-kube-api-access-lts6p\") pod \"horizon-operator-controller-manager-68c6d99b8f-8zwv5\" (UID: \"4881a1ed-a968-4f20-a8e6-94a91ec5eceb\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.367551 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.367978 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-bn5sn" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.388917 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.394035 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-h44mb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.413471 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.422394 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lts6p\" (UniqueName: \"kubernetes.io/projected/4881a1ed-a968-4f20-a8e6-94a91ec5eceb-kube-api-access-lts6p\") pod \"horizon-operator-controller-manager-68c6d99b8f-8zwv5\" (UID: \"4881a1ed-a968-4f20-a8e6-94a91ec5eceb\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.437206 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxc7\" (UniqueName: \"kubernetes.io/projected/a8fb4bb3-4b67-4ada-8876-389bff0e9836-kube-api-access-cnxc7\") pod \"heat-operator-controller-manager-5f64f6f8bb-xt2lq\" (UID: \"a8fb4bb3-4b67-4ada-8876-389bff0e9836\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.468464 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp4b2\" (UniqueName: \"kubernetes.io/projected/9106e41f-5788-46a2-86ce-6362832d6df7-kube-api-access-vp4b2\") pod \"manila-operator-controller-manager-7c79b5df47-qt45w\" (UID: \"9106e41f-5788-46a2-86ce-6362832d6df7\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.468528 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bct9w\" (UniqueName: \"kubernetes.io/projected/96da719f-5659-45eb-ac48-37165c936527-kube-api-access-bct9w\") pod \"ironic-operator-controller-manager-6c548fd776-79xxb\" (UID: \"96da719f-5659-45eb-ac48-37165c936527\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.468551 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndr62\" (UniqueName: \"kubernetes.io/projected/e944cb60-4e1a-448f-ab25-60ff87e3a166-kube-api-access-ndr62\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vkrn9\" (UID: \"e944cb60-4e1a-448f-ab25-60ff87e3a166\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.468620 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj7qp\" (UniqueName: \"kubernetes.io/projected/43c2b7af-9659-4ae6-8809-27475b31c611-kube-api-access-dj7qp\") pod \"keystone-operator-controller-manager-7765d96ddf-256rj\" (UID: \"43c2b7af-9659-4ae6-8809-27475b31c611\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.468644 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.468663 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6h2s\" (UniqueName: \"kubernetes.io/projected/0e880282-b396-4fa9-a3ac-89d81315ecd8-kube-api-access-h6h2s\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.469094 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:55:54 crc kubenswrapper[4990]: E1203 12:55:54.469862 4990 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:54 crc kubenswrapper[4990]: E1203 12:55:54.469933 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert podName:0e880282-b396-4fa9-a3ac-89d81315ecd8 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:54.969908346 +0000 UTC m=+1103.111819565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert") pod "infra-operator-controller-manager-57548d458d-zhgnf" (UID: "0e880282-b396-4fa9-a3ac-89d81315ecd8") : secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.491172 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.502132 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bct9w\" (UniqueName: \"kubernetes.io/projected/96da719f-5659-45eb-ac48-37165c936527-kube-api-access-bct9w\") pod \"ironic-operator-controller-manager-6c548fd776-79xxb\" (UID: \"96da719f-5659-45eb-ac48-37165c936527\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.531843 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.533667 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.539640 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.545392 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6h2s\" (UniqueName: \"kubernetes.io/projected/0e880282-b396-4fa9-a3ac-89d81315ecd8-kube-api-access-h6h2s\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.546226 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-56shl" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.567525 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.568878 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.576056 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp4b2\" (UniqueName: \"kubernetes.io/projected/9106e41f-5788-46a2-86ce-6362832d6df7-kube-api-access-vp4b2\") pod \"manila-operator-controller-manager-7c79b5df47-qt45w\" (UID: \"9106e41f-5788-46a2-86ce-6362832d6df7\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.576137 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndr62\" (UniqueName: \"kubernetes.io/projected/e944cb60-4e1a-448f-ab25-60ff87e3a166-kube-api-access-ndr62\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vkrn9\" (UID: \"e944cb60-4e1a-448f-ab25-60ff87e3a166\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.576185 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls59c\" (UniqueName: \"kubernetes.io/projected/a18809dc-c194-44b4-84e8-26e6da515bb7-kube-api-access-ls59c\") pod \"nova-operator-controller-manager-697bc559fc-dg8m2\" (UID: \"a18809dc-c194-44b4-84e8-26e6da515bb7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.576210 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgvjn\" (UniqueName: \"kubernetes.io/projected/0a2be3dd-196c-4186-ac5d-85ce550aa201-kube-api-access-hgvjn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-c8cgl\" (UID: \"0a2be3dd-196c-4186-ac5d-85ce550aa201\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.576235 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj7qp\" (UniqueName: \"kubernetes.io/projected/43c2b7af-9659-4ae6-8809-27475b31c611-kube-api-access-dj7qp\") pod \"keystone-operator-controller-manager-7765d96ddf-256rj\" (UID: \"43c2b7af-9659-4ae6-8809-27475b31c611\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.585397 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-blgsb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.620996 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-xg94h"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.627827 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndr62\" (UniqueName: \"kubernetes.io/projected/e944cb60-4e1a-448f-ab25-60ff87e3a166-kube-api-access-ndr62\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vkrn9\" (UID: \"e944cb60-4e1a-448f-ab25-60ff87e3a166\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.634605 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp4b2\" (UniqueName: \"kubernetes.io/projected/9106e41f-5788-46a2-86ce-6362832d6df7-kube-api-access-vp4b2\") pod \"manila-operator-controller-manager-7c79b5df47-qt45w\" (UID: \"9106e41f-5788-46a2-86ce-6362832d6df7\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.653173 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.653233 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-xg94h"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.653373 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.669343 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj7qp\" (UniqueName: \"kubernetes.io/projected/43c2b7af-9659-4ae6-8809-27475b31c611-kube-api-access-dj7qp\") pod \"keystone-operator-controller-manager-7765d96ddf-256rj\" (UID: \"43c2b7af-9659-4ae6-8809-27475b31c611\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.672792 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-jf7s9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.673394 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.677537 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls59c\" (UniqueName: \"kubernetes.io/projected/a18809dc-c194-44b4-84e8-26e6da515bb7-kube-api-access-ls59c\") pod \"nova-operator-controller-manager-697bc559fc-dg8m2\" (UID: \"a18809dc-c194-44b4-84e8-26e6da515bb7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.686224 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgvjn\" (UniqueName: \"kubernetes.io/projected/0a2be3dd-196c-4186-ac5d-85ce550aa201-kube-api-access-hgvjn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-c8cgl\" (UID: \"0a2be3dd-196c-4186-ac5d-85ce550aa201\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.686411 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27wbp\" (UniqueName: \"kubernetes.io/projected/6ca15749-37c2-477f-8a9a-33c07b97aaf3-kube-api-access-27wbp\") pod \"octavia-operator-controller-manager-998648c74-xg94h\" (UID: \"6ca15749-37c2-477f-8a9a-33c07b97aaf3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.679629 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.753242 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.770669 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.771643 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.788735 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.790783 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27wbp\" (UniqueName: \"kubernetes.io/projected/6ca15749-37c2-477f-8a9a-33c07b97aaf3-kube-api-access-27wbp\") pod \"octavia-operator-controller-manager-998648c74-xg94h\" (UID: \"6ca15749-37c2-477f-8a9a-33c07b97aaf3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.904298 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27wbp\" (UniqueName: \"kubernetes.io/projected/6ca15749-37c2-477f-8a9a-33c07b97aaf3-kube-api-access-27wbp\") pod \"octavia-operator-controller-manager-998648c74-xg94h\" (UID: \"6ca15749-37c2-477f-8a9a-33c07b97aaf3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.908174 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x"] Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.911358 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.912081 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.934102 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.934522 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nspl8" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.936241 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rkqdf" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.950382 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgvjn\" (UniqueName: \"kubernetes.io/projected/0a2be3dd-196c-4186-ac5d-85ce550aa201-kube-api-access-hgvjn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-c8cgl\" (UID: \"0a2be3dd-196c-4186-ac5d-85ce550aa201\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:55:54 crc kubenswrapper[4990]: I1203 12:55:54.984802 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.003300 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj9t7\" (UniqueName: \"kubernetes.io/projected/c0833459-1161-4beb-ad68-07d51b5b33d7-kube-api-access-vj9t7\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.003380 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.004731 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x"] Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.004935 4990 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.004982 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert podName:0e880282-b396-4fa9-a3ac-89d81315ecd8 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:56.004968314 +0000 UTC m=+1104.146879543 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert") pod "infra-operator-controller-manager-57548d458d-zhgnf" (UID: "0e880282-b396-4fa9-a3ac-89d81315ecd8") : secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.004858 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.017971 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.027544 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.028299 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls59c\" (UniqueName: \"kubernetes.io/projected/a18809dc-c194-44b4-84e8-26e6da515bb7-kube-api-access-ls59c\") pod \"nova-operator-controller-manager-697bc559fc-dg8m2\" (UID: \"a18809dc-c194-44b4-84e8-26e6da515bb7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.031707 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-48fht" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.036343 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.054519 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.055977 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.059278 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.061056 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.061369 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-q6m6b" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.064602 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.062992 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-qcv5r" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.069720 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.085897 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.108145 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drxl5\" (UniqueName: \"kubernetes.io/projected/64f072f3-940c-431d-8e0f-b77e9349e79e-kube-api-access-drxl5\") pod \"ovn-operator-controller-manager-b6456fdb6-4db2x\" (UID: \"64f072f3-940c-431d-8e0f-b77e9349e79e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.108216 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.108350 4990 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.108661 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert podName:c0833459-1161-4beb-ad68-07d51b5b33d7 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:55.608632687 +0000 UTC m=+1103.750543916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" (UID: "c0833459-1161-4beb-ad68-07d51b5b33d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.108927 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj9t7\" (UniqueName: \"kubernetes.io/projected/c0833459-1161-4beb-ad68-07d51b5b33d7-kube-api-access-vj9t7\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.114047 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.115525 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.117496 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-79l6k" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.147540 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.147610 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj9t7\" (UniqueName: \"kubernetes.io/projected/c0833459-1161-4beb-ad68-07d51b5b33d7-kube-api-access-vj9t7\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.152730 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.158791 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-8b8f5" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.168997 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.198344 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.209853 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tq9g\" (UniqueName: \"kubernetes.io/projected/8bf181cd-d802-43c1-a7f0-7df5ec5008a1-kube-api-access-7tq9g\") pod \"swift-operator-controller-manager-5f8c65bbfc-rvslq\" (UID: \"8bf181cd-d802-43c1-a7f0-7df5ec5008a1\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.209908 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrsvd\" (UniqueName: \"kubernetes.io/projected/3443975f-d482-422a-a010-9940c2b1182f-kube-api-access-wrsvd\") pod \"placement-operator-controller-manager-78f8948974-x7jtx\" (UID: \"3443975f-d482-422a-a010-9940c2b1182f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.209975 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drxl5\" (UniqueName: \"kubernetes.io/projected/64f072f3-940c-431d-8e0f-b77e9349e79e-kube-api-access-drxl5\") pod \"ovn-operator-controller-manager-b6456fdb6-4db2x\" (UID: \"64f072f3-940c-431d-8e0f-b77e9349e79e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.210052 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fltvm\" (UniqueName: \"kubernetes.io/projected/213fe9a3-0270-4ec5-b380-2f28159b5b6a-kube-api-access-fltvm\") pod \"telemetry-operator-controller-manager-76cc84c6bb-4wkxw\" (UID: \"213fe9a3-0270-4ec5-b380-2f28159b5b6a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.230615 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drxl5\" (UniqueName: \"kubernetes.io/projected/64f072f3-940c-431d-8e0f-b77e9349e79e-kube-api-access-drxl5\") pod \"ovn-operator-controller-manager-b6456fdb6-4db2x\" (UID: \"64f072f3-940c-431d-8e0f-b77e9349e79e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.246611 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.247946 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.249045 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.253291 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.253617 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-tzbjv" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.255397 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.269545 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.300689 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.317826 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fltvm\" (UniqueName: \"kubernetes.io/projected/213fe9a3-0270-4ec5-b380-2f28159b5b6a-kube-api-access-fltvm\") pod \"telemetry-operator-controller-manager-76cc84c6bb-4wkxw\" (UID: \"213fe9a3-0270-4ec5-b380-2f28159b5b6a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.317930 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tq9g\" (UniqueName: \"kubernetes.io/projected/8bf181cd-d802-43c1-a7f0-7df5ec5008a1-kube-api-access-7tq9g\") pod \"swift-operator-controller-manager-5f8c65bbfc-rvslq\" (UID: \"8bf181cd-d802-43c1-a7f0-7df5ec5008a1\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.317977 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6mkp\" (UniqueName: \"kubernetes.io/projected/425132be-e04e-46f6-ac06-1546c5b12a26-kube-api-access-g6mkp\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.318006 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrsvd\" (UniqueName: \"kubernetes.io/projected/3443975f-d482-422a-a010-9940c2b1182f-kube-api-access-wrsvd\") pod \"placement-operator-controller-manager-78f8948974-x7jtx\" (UID: \"3443975f-d482-422a-a010-9940c2b1182f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.318065 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.318124 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgbg9\" (UniqueName: \"kubernetes.io/projected/4481645b-1e70-415a-a01b-131b1761dd19-kube-api-access-xgbg9\") pod \"test-operator-controller-manager-5854674fcc-tg4ww\" (UID: \"4481645b-1e70-415a-a01b-131b1761dd19\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.318166 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcmxg\" (UniqueName: \"kubernetes.io/projected/d2c9d382-f60c-4c2c-a684-ddec4371b165-kube-api-access-xcmxg\") pod \"watcher-operator-controller-manager-769dc69bc-7cl8r\" (UID: \"d2c9d382-f60c-4c2c-a684-ddec4371b165\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.318266 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.334106 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.335364 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.341265 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-crktk" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.356806 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrsvd\" (UniqueName: \"kubernetes.io/projected/3443975f-d482-422a-a010-9940c2b1182f-kube-api-access-wrsvd\") pod \"placement-operator-controller-manager-78f8948974-x7jtx\" (UID: \"3443975f-d482-422a-a010-9940c2b1182f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.359811 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.363647 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tq9g\" (UniqueName: \"kubernetes.io/projected/8bf181cd-d802-43c1-a7f0-7df5ec5008a1-kube-api-access-7tq9g\") pod \"swift-operator-controller-manager-5f8c65bbfc-rvslq\" (UID: \"8bf181cd-d802-43c1-a7f0-7df5ec5008a1\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.366593 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fltvm\" (UniqueName: \"kubernetes.io/projected/213fe9a3-0270-4ec5-b380-2f28159b5b6a-kube-api-access-fltvm\") pod \"telemetry-operator-controller-manager-76cc84c6bb-4wkxw\" (UID: \"213fe9a3-0270-4ec5-b380-2f28159b5b6a\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.405666 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.433239 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6mkp\" (UniqueName: \"kubernetes.io/projected/425132be-e04e-46f6-ac06-1546c5b12a26-kube-api-access-g6mkp\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.433296 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.433361 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgbg9\" (UniqueName: \"kubernetes.io/projected/4481645b-1e70-415a-a01b-131b1761dd19-kube-api-access-xgbg9\") pod \"test-operator-controller-manager-5854674fcc-tg4ww\" (UID: \"4481645b-1e70-415a-a01b-131b1761dd19\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.433393 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcmxg\" (UniqueName: \"kubernetes.io/projected/d2c9d382-f60c-4c2c-a684-ddec4371b165-kube-api-access-xcmxg\") pod \"watcher-operator-controller-manager-769dc69bc-7cl8r\" (UID: \"d2c9d382-f60c-4c2c-a684-ddec4371b165\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.433427 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.433527 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkxhd\" (UniqueName: \"kubernetes.io/projected/cc2d7750-b3b5-4585-95fd-4f84e1c437ad-kube-api-access-fkxhd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-p8hpt\" (UID: \"cc2d7750-b3b5-4585-95fd-4f84e1c437ad\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.433969 4990 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.434011 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:55.933998296 +0000 UTC m=+1104.075909525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "metrics-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.434406 4990 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.434430 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:55.934422368 +0000 UTC m=+1104.076333597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.456479 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgbg9\" (UniqueName: \"kubernetes.io/projected/4481645b-1e70-415a-a01b-131b1761dd19-kube-api-access-xgbg9\") pod \"test-operator-controller-manager-5854674fcc-tg4ww\" (UID: \"4481645b-1e70-415a-a01b-131b1761dd19\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.456763 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.463796 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.481959 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6mkp\" (UniqueName: \"kubernetes.io/projected/425132be-e04e-46f6-ac06-1546c5b12a26-kube-api-access-g6mkp\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.482805 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.499314 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" event={"ID":"c7a7a3b2-d356-4dfa-b640-ae39aa2d4840","Type":"ContainerStarted","Data":"2ec3b257b6bcc1790453e5335eb446cbb9e399c6be028912a39923753eaf5969"} Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.500406 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" event={"ID":"c45d3b36-d188-4bb1-b19f-5521709f572a","Type":"ContainerStarted","Data":"1f4ac975dc9bc975338fad2ebe86d4b2b3c1d4602edf2986cefdffc53baae02e"} Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.524356 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcmxg\" (UniqueName: \"kubernetes.io/projected/d2c9d382-f60c-4c2c-a684-ddec4371b165-kube-api-access-xcmxg\") pod \"watcher-operator-controller-manager-769dc69bc-7cl8r\" (UID: \"d2c9d382-f60c-4c2c-a684-ddec4371b165\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.534574 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkxhd\" (UniqueName: \"kubernetes.io/projected/cc2d7750-b3b5-4585-95fd-4f84e1c437ad-kube-api-access-fkxhd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-p8hpt\" (UID: \"cc2d7750-b3b5-4585-95fd-4f84e1c437ad\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.547542 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.561934 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkxhd\" (UniqueName: \"kubernetes.io/projected/cc2d7750-b3b5-4585-95fd-4f84e1c437ad-kube-api-access-fkxhd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-p8hpt\" (UID: \"cc2d7750-b3b5-4585-95fd-4f84e1c437ad\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.563834 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.590647 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.616491 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.639671 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.641942 4990 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.642046 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert podName:c0833459-1161-4beb-ad68-07d51b5b33d7 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:56.642009441 +0000 UTC m=+1104.783920680 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" (UID: "c0833459-1161-4beb-ad68-07d51b5b33d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.696585 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.846061 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc"] Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.870213 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w"] Dec 03 12:55:55 crc kubenswrapper[4990]: W1203 12:55:55.875088 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac715eff_4f7f_44b0_b5ca_c5d283edca36.slice/crio-01ef873f4998e9600806a26af56862be9cea92a748be94cf1b170839e8b897a6 WatchSource:0}: Error finding container 01ef873f4998e9600806a26af56862be9cea92a748be94cf1b170839e8b897a6: Status 404 returned error can't find the container with id 01ef873f4998e9600806a26af56862be9cea92a748be94cf1b170839e8b897a6 Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.947110 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: I1203 12:55:55.947248 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.947423 4990 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.947491 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:56.947472675 +0000 UTC m=+1105.089383904 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "metrics-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.947637 4990 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 12:55:55 crc kubenswrapper[4990]: E1203 12:55:55.947721 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:56.947696801 +0000 UTC m=+1105.089608090 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "webhook-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.053026 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.053239 4990 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.053338 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert podName:0e880282-b396-4fa9-a3ac-89d81315ecd8 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:58.053319997 +0000 UTC m=+1106.195231226 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert") pod "infra-operator-controller-manager-57548d458d-zhgnf" (UID: "0e880282-b396-4fa9-a3ac-89d81315ecd8") : secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.251763 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.341166 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.341235 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.341253 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.341269 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb"] Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.351197 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda18809dc_c194_44b4_84e8_26e6da515bb7.slice/crio-466b27521e09f90f81a2a45ef180c32d52846d71a882042a3e2f74bfa98315f5 WatchSource:0}: Error finding container 466b27521e09f90f81a2a45ef180c32d52846d71a882042a3e2f74bfa98315f5: Status 404 returned error can't find the container with id 466b27521e09f90f81a2a45ef180c32d52846d71a882042a3e2f74bfa98315f5 Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.353833 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.367661 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-xg94h"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.370860 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.481885 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.490793 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl"] Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.515601 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" event={"ID":"a18809dc-c194-44b4-84e8-26e6da515bb7","Type":"ContainerStarted","Data":"466b27521e09f90f81a2a45ef180c32d52846d71a882042a3e2f74bfa98315f5"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.517667 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" event={"ID":"96da719f-5659-45eb-ac48-37165c936527","Type":"ContainerStarted","Data":"22bad89f290522408e31bbb84681e4fb8566acd6693bfba1cd27752992f4069a"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.518918 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" event={"ID":"ac715eff-4f7f-44b0-b5ca-c5d283edca36","Type":"ContainerStarted","Data":"01ef873f4998e9600806a26af56862be9cea92a748be94cf1b170839e8b897a6"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.519933 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" event={"ID":"64f072f3-940c-431d-8e0f-b77e9349e79e","Type":"ContainerStarted","Data":"07485a55f11bb8855eaa31af058b7cb8e63734f181d4c0aa0094e99632695faf"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.521706 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" event={"ID":"43c2b7af-9659-4ae6-8809-27475b31c611","Type":"ContainerStarted","Data":"a2701e483d5cac4e190e135601c9a19f021d5774c81e43661a93298e6fc3b62d"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.522593 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" event={"ID":"93f0e562-5973-4c2a-ab48-aa167ae49ffa","Type":"ContainerStarted","Data":"bc267b553f0d4500ae55dc7e7b3f02d8cfa852bce45cdd22a5562cc6991c6341"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.523748 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" event={"ID":"4881a1ed-a968-4f20-a8e6-94a91ec5eceb","Type":"ContainerStarted","Data":"4946a8389bec0c4a83751931adf0595aa263f6745fe2d117e9d8695f4f3fceef"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.525120 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" event={"ID":"e944cb60-4e1a-448f-ab25-60ff87e3a166","Type":"ContainerStarted","Data":"433c74c144a2110b2a1128f654c0e98155caf1effcbaec6754a332fa64199c07"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.530041 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" event={"ID":"a8fb4bb3-4b67-4ada-8876-389bff0e9836","Type":"ContainerStarted","Data":"77b25c44c3b6d84ca64ecf97340732d1077618e8d28754258bdbf44e6773bc6d"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.533320 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" event={"ID":"6ca15749-37c2-477f-8a9a-33c07b97aaf3","Type":"ContainerStarted","Data":"8a5cea64220d8bc3335e65ac3abec3e3a97a6a5e398d78260775c426b71a1ea7"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.534570 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" event={"ID":"9106e41f-5788-46a2-86ce-6362832d6df7","Type":"ContainerStarted","Data":"310b4c5632258d8503347853249eaf22ed578e8026b6930f6a1551cbc4537500"} Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.558145 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw"] Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.566083 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod213fe9a3_0270_4ec5_b380_2f28159b5b6a.slice/crio-4da0dad7c2195e4c7bbf208d41f2ca972eff4a615dc2f55d19995fe15ea788d4 WatchSource:0}: Error finding container 4da0dad7c2195e4c7bbf208d41f2ca972eff4a615dc2f55d19995fe15ea788d4: Status 404 returned error can't find the container with id 4da0dad7c2195e4c7bbf208d41f2ca972eff4a615dc2f55d19995fe15ea788d4 Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.567178 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx"] Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.574842 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fltvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-4wkxw_openstack-operators(213fe9a3-0270-4ec5-b380-2f28159b5b6a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.578724 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fltvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-4wkxw_openstack-operators(213fe9a3-0270-4ec5-b380-2f28159b5b6a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.581340 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" podUID="213fe9a3-0270-4ec5-b380-2f28159b5b6a" Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.581868 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww"] Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.582046 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3443975f_d482_422a_a010_9940c2b1182f.slice/crio-9b5065a5dd923fd73314a195420132e476d3fe652db0416bf2f36e30c0fa31a3 WatchSource:0}: Error finding container 9b5065a5dd923fd73314a195420132e476d3fe652db0416bf2f36e30c0fa31a3: Status 404 returned error can't find the container with id 9b5065a5dd923fd73314a195420132e476d3fe652db0416bf2f36e30c0fa31a3 Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.584841 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wrsvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-x7jtx_openstack-operators(3443975f-d482-422a-a010-9940c2b1182f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.585208 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4481645b_1e70_415a_a01b_131b1761dd19.slice/crio-d036ddfcb854263ece3d5b1e79f08f029c8cc467ac996ff70591f36a6f957b6a WatchSource:0}: Error finding container d036ddfcb854263ece3d5b1e79f08f029c8cc467ac996ff70591f36a6f957b6a: Status 404 returned error can't find the container with id d036ddfcb854263ece3d5b1e79f08f029c8cc467ac996ff70591f36a6f957b6a Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.589130 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq"] Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.589291 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bf181cd_d802_43c1_a7f0_7df5ec5008a1.slice/crio-27fe0de4a1c23b22ea9f65f9115db4a55ce495120d5dbba0c45d1ff03c99452e WatchSource:0}: Error finding container 27fe0de4a1c23b22ea9f65f9115db4a55ce495120d5dbba0c45d1ff03c99452e: Status 404 returned error can't find the container with id 27fe0de4a1c23b22ea9f65f9115db4a55ce495120d5dbba0c45d1ff03c99452e Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.591430 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xgbg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tg4ww_openstack-operators(4481645b-1e70-415a-a01b-131b1761dd19): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.594604 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tq9g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-rvslq_openstack-operators(8bf181cd-d802-43c1-a7f0-7df5ec5008a1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.596822 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r"] Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.597731 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2c9d382_f60c_4c2c_a684_ddec4371b165.slice/crio-b8be4c90570930c3f81168dd0f08a6c6c5fd95ce1f150281dd2cb0d1aa5678c1 WatchSource:0}: Error finding container b8be4c90570930c3f81168dd0f08a6c6c5fd95ce1f150281dd2cb0d1aa5678c1: Status 404 returned error can't find the container with id b8be4c90570930c3f81168dd0f08a6c6c5fd95ce1f150281dd2cb0d1aa5678c1 Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.597764 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tq9g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-rvslq_openstack-operators(8bf181cd-d802-43c1-a7f0-7df5ec5008a1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.597862 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xgbg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tg4ww_openstack-operators(4481645b-1e70-415a-a01b-131b1761dd19): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.599143 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" podUID="8bf181cd-d802-43c1-a7f0-7df5ec5008a1" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.599203 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" podUID="4481645b-1e70-415a-a01b-131b1761dd19" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.601795 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xcmxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-7cl8r_openstack-operators(d2c9d382-f60c-4c2c-a684-ddec4371b165): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: W1203 12:55:56.603343 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc2d7750_b3b5_4585_95fd_4f84e1c437ad.slice/crio-27ed279f25d7d50d7de1aa2f1944945cf829cf5c3fbe2367db2064a70df6c69f WatchSource:0}: Error finding container 27ed279f25d7d50d7de1aa2f1944945cf829cf5c3fbe2367db2064a70df6c69f: Status 404 returned error can't find the container with id 27ed279f25d7d50d7de1aa2f1944945cf829cf5c3fbe2367db2064a70df6c69f Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.603436 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt"] Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.606901 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xcmxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-7cl8r_openstack-operators(d2c9d382-f60c-4c2c-a684-ddec4371b165): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.608265 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" podUID="d2c9d382-f60c-4c2c-a684-ddec4371b165" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.609358 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fkxhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-p8hpt_openstack-operators(cc2d7750-b3b5-4585-95fd-4f84e1c437ad): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.610777 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" podUID="cc2d7750-b3b5-4585-95fd-4f84e1c437ad" Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.673398 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.673603 4990 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.673704 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert podName:c0833459-1161-4beb-ad68-07d51b5b33d7 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:58.673681112 +0000 UTC m=+1106.815592341 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" (UID: "c0833459-1161-4beb-ad68-07d51b5b33d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.976932 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:56 crc kubenswrapper[4990]: I1203 12:55:56.977046 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.977135 4990 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.977237 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:58.977215702 +0000 UTC m=+1107.119126931 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "metrics-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.977248 4990 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 12:55:56 crc kubenswrapper[4990]: E1203 12:55:56.977321 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:55:58.977300315 +0000 UTC m=+1107.119211544 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "webhook-server-cert" not found Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.544263 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" event={"ID":"0a2be3dd-196c-4186-ac5d-85ce550aa201","Type":"ContainerStarted","Data":"8f57436153aaa3f689f22ab80bab33b0e75b28c3f5e81eaa3717cc0e18432664"} Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.546237 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" event={"ID":"cc2d7750-b3b5-4585-95fd-4f84e1c437ad","Type":"ContainerStarted","Data":"27ed279f25d7d50d7de1aa2f1944945cf829cf5c3fbe2367db2064a70df6c69f"} Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.547550 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" event={"ID":"8bf181cd-d802-43c1-a7f0-7df5ec5008a1","Type":"ContainerStarted","Data":"27fe0de4a1c23b22ea9f65f9115db4a55ce495120d5dbba0c45d1ff03c99452e"} Dec 03 12:55:57 crc kubenswrapper[4990]: E1203 12:55:57.550598 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" podUID="cc2d7750-b3b5-4585-95fd-4f84e1c437ad" Dec 03 12:55:57 crc kubenswrapper[4990]: E1203 12:55:57.551361 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" podUID="8bf181cd-d802-43c1-a7f0-7df5ec5008a1" Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.551541 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" event={"ID":"3443975f-d482-422a-a010-9940c2b1182f","Type":"ContainerStarted","Data":"9b5065a5dd923fd73314a195420132e476d3fe652db0416bf2f36e30c0fa31a3"} Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.553950 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" event={"ID":"d2c9d382-f60c-4c2c-a684-ddec4371b165","Type":"ContainerStarted","Data":"b8be4c90570930c3f81168dd0f08a6c6c5fd95ce1f150281dd2cb0d1aa5678c1"} Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.558634 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" event={"ID":"213fe9a3-0270-4ec5-b380-2f28159b5b6a","Type":"ContainerStarted","Data":"4da0dad7c2195e4c7bbf208d41f2ca972eff4a615dc2f55d19995fe15ea788d4"} Dec 03 12:55:57 crc kubenswrapper[4990]: I1203 12:55:57.561072 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" event={"ID":"4481645b-1e70-415a-a01b-131b1761dd19","Type":"ContainerStarted","Data":"d036ddfcb854263ece3d5b1e79f08f029c8cc467ac996ff70591f36a6f957b6a"} Dec 03 12:55:57 crc kubenswrapper[4990]: E1203 12:55:57.561589 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" podUID="213fe9a3-0270-4ec5-b380-2f28159b5b6a" Dec 03 12:55:57 crc kubenswrapper[4990]: E1203 12:55:57.561656 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" podUID="d2c9d382-f60c-4c2c-a684-ddec4371b165" Dec 03 12:55:57 crc kubenswrapper[4990]: E1203 12:55:57.565680 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" podUID="4481645b-1e70-415a-a01b-131b1761dd19" Dec 03 12:55:58 crc kubenswrapper[4990]: I1203 12:55:58.094191 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.094371 4990 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.094420 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert podName:0e880282-b396-4fa9-a3ac-89d81315ecd8 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:02.094406306 +0000 UTC m=+1110.236317525 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert") pod "infra-operator-controller-manager-57548d458d-zhgnf" (UID: "0e880282-b396-4fa9-a3ac-89d81315ecd8") : secret "infra-operator-webhook-server-cert" not found Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.576975 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" podUID="213fe9a3-0270-4ec5-b380-2f28159b5b6a" Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.577019 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" podUID="cc2d7750-b3b5-4585-95fd-4f84e1c437ad" Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.577072 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" podUID="4481645b-1e70-415a-a01b-131b1761dd19" Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.577768 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" podUID="8bf181cd-d802-43c1-a7f0-7df5ec5008a1" Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.577945 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" podUID="d2c9d382-f60c-4c2c-a684-ddec4371b165" Dec 03 12:55:58 crc kubenswrapper[4990]: I1203 12:55:58.709684 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.709914 4990 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:58 crc kubenswrapper[4990]: E1203 12:55:58.710012 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert podName:c0833459-1161-4beb-ad68-07d51b5b33d7 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:02.709991431 +0000 UTC m=+1110.851902660 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" (UID: "c0833459-1161-4beb-ad68-07d51b5b33d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:55:59 crc kubenswrapper[4990]: I1203 12:55:59.015797 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:59 crc kubenswrapper[4990]: I1203 12:55:59.015969 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:55:59 crc kubenswrapper[4990]: E1203 12:55:59.016047 4990 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 12:55:59 crc kubenswrapper[4990]: E1203 12:55:59.016155 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:03.01612841 +0000 UTC m=+1111.158039829 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "metrics-server-cert" not found Dec 03 12:55:59 crc kubenswrapper[4990]: E1203 12:55:59.016162 4990 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 12:55:59 crc kubenswrapper[4990]: E1203 12:55:59.016246 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:03.016219143 +0000 UTC m=+1111.158130372 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "webhook-server-cert" not found Dec 03 12:56:02 crc kubenswrapper[4990]: I1203 12:56:02.175176 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:56:02 crc kubenswrapper[4990]: E1203 12:56:02.175426 4990 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 12:56:02 crc kubenswrapper[4990]: E1203 12:56:02.176434 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert podName:0e880282-b396-4fa9-a3ac-89d81315ecd8 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:10.176404371 +0000 UTC m=+1118.318315600 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert") pod "infra-operator-controller-manager-57548d458d-zhgnf" (UID: "0e880282-b396-4fa9-a3ac-89d81315ecd8") : secret "infra-operator-webhook-server-cert" not found Dec 03 12:56:02 crc kubenswrapper[4990]: I1203 12:56:02.783591 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:56:02 crc kubenswrapper[4990]: E1203 12:56:02.784136 4990 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:56:02 crc kubenswrapper[4990]: E1203 12:56:02.784276 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert podName:c0833459-1161-4beb-ad68-07d51b5b33d7 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:10.784257233 +0000 UTC m=+1118.926168462 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" (UID: "c0833459-1161-4beb-ad68-07d51b5b33d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 12:56:03 crc kubenswrapper[4990]: I1203 12:56:03.088358 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:03 crc kubenswrapper[4990]: E1203 12:56:03.088546 4990 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 12:56:03 crc kubenswrapper[4990]: I1203 12:56:03.088852 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:03 crc kubenswrapper[4990]: E1203 12:56:03.088896 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:11.088875962 +0000 UTC m=+1119.230787191 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "webhook-server-cert" not found Dec 03 12:56:03 crc kubenswrapper[4990]: E1203 12:56:03.088943 4990 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 12:56:03 crc kubenswrapper[4990]: E1203 12:56:03.088982 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs podName:425132be-e04e-46f6-ac06-1546c5b12a26 nodeName:}" failed. No retries permitted until 2025-12-03 12:56:11.088968645 +0000 UTC m=+1119.230879874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs") pod "openstack-operator-controller-manager-5b474d9948-p5wjj" (UID: "425132be-e04e-46f6-ac06-1546c5b12a26") : secret "metrics-server-cert" not found Dec 03 12:56:03 crc kubenswrapper[4990]: I1203 12:56:03.286423 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:56:03 crc kubenswrapper[4990]: I1203 12:56:03.286596 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:56:10 crc kubenswrapper[4990]: I1203 12:56:10.223565 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:56:10 crc kubenswrapper[4990]: I1203 12:56:10.235342 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e880282-b396-4fa9-a3ac-89d81315ecd8-cert\") pod \"infra-operator-controller-manager-57548d458d-zhgnf\" (UID: \"0e880282-b396-4fa9-a3ac-89d81315ecd8\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:56:10 crc kubenswrapper[4990]: I1203 12:56:10.408910 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:56:10 crc kubenswrapper[4990]: I1203 12:56:10.835302 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:56:10 crc kubenswrapper[4990]: I1203 12:56:10.842329 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0833459-1161-4beb-ad68-07d51b5b33d7-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9\" (UID: \"c0833459-1161-4beb-ad68-07d51b5b33d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:56:11 crc kubenswrapper[4990]: I1203 12:56:11.050871 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:56:11 crc kubenswrapper[4990]: I1203 12:56:11.140313 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:11 crc kubenswrapper[4990]: I1203 12:56:11.140428 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:11 crc kubenswrapper[4990]: I1203 12:56:11.144018 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-webhook-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:11 crc kubenswrapper[4990]: I1203 12:56:11.144018 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/425132be-e04e-46f6-ac06-1546c5b12a26-metrics-certs\") pod \"openstack-operator-controller-manager-5b474d9948-p5wjj\" (UID: \"425132be-e04e-46f6-ac06-1546c5b12a26\") " pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:11 crc kubenswrapper[4990]: I1203 12:56:11.275488 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:12 crc kubenswrapper[4990]: E1203 12:56:12.157975 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 03 12:56:12 crc kubenswrapper[4990]: E1203 12:56:12.158169 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-27wbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-xg94h_openstack-operators(6ca15749-37c2-477f-8a9a-33c07b97aaf3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:13 crc kubenswrapper[4990]: E1203 12:56:13.713578 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 12:56:13 crc kubenswrapper[4990]: E1203 12:56:13.714656 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dj7qp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-256rj_openstack-operators(43c2b7af-9659-4ae6-8809-27475b31c611): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:14 crc kubenswrapper[4990]: E1203 12:56:14.355375 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 12:56:14 crc kubenswrapper[4990]: E1203 12:56:14.355733 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ls59c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-dg8m2_openstack-operators(a18809dc-c194-44b4-84e8-26e6da515bb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:25 crc kubenswrapper[4990]: E1203 12:56:25.551738 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 03 12:56:25 crc kubenswrapper[4990]: E1203 12:56:25.555474 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xgbg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tg4ww_openstack-operators(4481645b-1e70-415a-a01b-131b1761dd19): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:25 crc kubenswrapper[4990]: I1203 12:56:25.559236 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 12:56:27 crc kubenswrapper[4990]: E1203 12:56:27.435954 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 03 12:56:27 crc kubenswrapper[4990]: E1203 12:56:27.436840 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tq9g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-rvslq_openstack-operators(8bf181cd-d802-43c1-a7f0-7df5ec5008a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:28 crc kubenswrapper[4990]: E1203 12:56:28.941728 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 03 12:56:28 crc kubenswrapper[4990]: E1203 12:56:28.942209 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xcmxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-7cl8r_openstack-operators(d2c9d382-f60c-4c2c-a684-ddec4371b165): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:29 crc kubenswrapper[4990]: I1203 12:56:29.942539 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj"] Dec 03 12:56:29 crc kubenswrapper[4990]: I1203 12:56:29.995529 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf"] Dec 03 12:56:30 crc kubenswrapper[4990]: I1203 12:56:30.005117 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9"] Dec 03 12:56:30 crc kubenswrapper[4990]: W1203 12:56:30.442914 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod425132be_e04e_46f6_ac06_1546c5b12a26.slice/crio-ef60b387262dcdeb85e71beb3f3c37d095df12f9fd80e822a4cd07598ab06980 WatchSource:0}: Error finding container ef60b387262dcdeb85e71beb3f3c37d095df12f9fd80e822a4cd07598ab06980: Status 404 returned error can't find the container with id ef60b387262dcdeb85e71beb3f3c37d095df12f9fd80e822a4cd07598ab06980 Dec 03 12:56:30 crc kubenswrapper[4990]: W1203 12:56:30.443829 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e880282_b396_4fa9_a3ac_89d81315ecd8.slice/crio-c104f2d6fcf32dace606f1c0488e3497d14a8efb3ac9aedff143352c87b2cb20 WatchSource:0}: Error finding container c104f2d6fcf32dace606f1c0488e3497d14a8efb3ac9aedff143352c87b2cb20: Status 404 returned error can't find the container with id c104f2d6fcf32dace606f1c0488e3497d14a8efb3ac9aedff143352c87b2cb20 Dec 03 12:56:30 crc kubenswrapper[4990]: E1203 12:56:30.492571 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 12:56:30 crc kubenswrapper[4990]: E1203 12:56:30.492805 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fkxhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-p8hpt_openstack-operators(cc2d7750-b3b5-4585-95fd-4f84e1c437ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:30 crc kubenswrapper[4990]: E1203 12:56:30.494088 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" podUID="cc2d7750-b3b5-4585-95fd-4f84e1c437ad" Dec 03 12:56:30 crc kubenswrapper[4990]: I1203 12:56:30.839983 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" event={"ID":"0e880282-b396-4fa9-a3ac-89d81315ecd8","Type":"ContainerStarted","Data":"c104f2d6fcf32dace606f1c0488e3497d14a8efb3ac9aedff143352c87b2cb20"} Dec 03 12:56:30 crc kubenswrapper[4990]: I1203 12:56:30.841377 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" event={"ID":"c0833459-1161-4beb-ad68-07d51b5b33d7","Type":"ContainerStarted","Data":"b739b58016638f668caf9200fdc8ce43d738e51c026069364f0c084986f0b47a"} Dec 03 12:56:30 crc kubenswrapper[4990]: I1203 12:56:30.842934 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" event={"ID":"425132be-e04e-46f6-ac06-1546c5b12a26","Type":"ContainerStarted","Data":"ef60b387262dcdeb85e71beb3f3c37d095df12f9fd80e822a4cd07598ab06980"} Dec 03 12:56:31 crc kubenswrapper[4990]: I1203 12:56:31.856224 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" event={"ID":"93f0e562-5973-4c2a-ab48-aa167ae49ffa","Type":"ContainerStarted","Data":"3befd5f56de35f7c89da2ec1b5817ccf4f663667d6d82a3c9aad2c6ca457d05b"} Dec 03 12:56:31 crc kubenswrapper[4990]: I1203 12:56:31.858804 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" event={"ID":"96da719f-5659-45eb-ac48-37165c936527","Type":"ContainerStarted","Data":"96c40b348d36898c52f53f2de26e7e271e877578b37ac61ab4e6a016ab6ae0de"} Dec 03 12:56:32 crc kubenswrapper[4990]: E1203 12:56:32.383248 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 12:56:32 crc kubenswrapper[4990]: E1203 12:56:32.383863 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dj7qp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-256rj_openstack-operators(43c2b7af-9659-4ae6-8809-27475b31c611): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 12:56:32 crc kubenswrapper[4990]: E1203 12:56:32.385466 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" podUID="43c2b7af-9659-4ae6-8809-27475b31c611" Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.868057 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" event={"ID":"c7a7a3b2-d356-4dfa-b640-ae39aa2d4840","Type":"ContainerStarted","Data":"b7274a9cb4176a6813ce20a0f2b740c49e3b54ce5451dffeb57ac2876f18afbd"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.870158 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" event={"ID":"e944cb60-4e1a-448f-ab25-60ff87e3a166","Type":"ContainerStarted","Data":"e8272905fb9b8d0e6dcaacafffea4b52a1607680dc82607c4ed560e107cb603f"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.871920 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" event={"ID":"a8fb4bb3-4b67-4ada-8876-389bff0e9836","Type":"ContainerStarted","Data":"39ff723384164cf845a4f6311cf5a4777c02ee89f60e9305a953d253b767462b"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.873128 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" event={"ID":"4881a1ed-a968-4f20-a8e6-94a91ec5eceb","Type":"ContainerStarted","Data":"7dfe4b0e9b132fe19f6105888636df1dd86c99437c9897e5a6e465b1f136c55c"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.874522 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" event={"ID":"ac715eff-4f7f-44b0-b5ca-c5d283edca36","Type":"ContainerStarted","Data":"592dfccd2e086a6eef81b984f9fee1e4f43b91d4136c84e7d2d7db6cc0b71d8d"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.876490 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" event={"ID":"0a2be3dd-196c-4186-ac5d-85ce550aa201","Type":"ContainerStarted","Data":"3579819176e664d87e4d279353cecb3c607d9fb3f99b13c4a319f27b28e451c2"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.877973 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" event={"ID":"9106e41f-5788-46a2-86ce-6362832d6df7","Type":"ContainerStarted","Data":"f9c58e54cc345ea62b757503e3530053dcfb0d8fccd65aa1dd1a515783aef7cb"} Dec 03 12:56:32 crc kubenswrapper[4990]: I1203 12:56:32.879950 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" event={"ID":"c45d3b36-d188-4bb1-b19f-5521709f572a","Type":"ContainerStarted","Data":"c6f3af044d80c5420da9af21a8177b3e43fd416708093ec889e3d53b417482d0"} Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.287089 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.287160 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.287213 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.287904 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80c0112fe475b4f4e2b7d1d072f2e686f621ee22a4d45518867b967855b8112b"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.287967 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://80c0112fe475b4f4e2b7d1d072f2e686f621ee22a4d45518867b967855b8112b" gracePeriod=600 Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.889159 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="80c0112fe475b4f4e2b7d1d072f2e686f621ee22a4d45518867b967855b8112b" exitCode=0 Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.889207 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"80c0112fe475b4f4e2b7d1d072f2e686f621ee22a4d45518867b967855b8112b"} Dec 03 12:56:33 crc kubenswrapper[4990]: I1203 12:56:33.889247 4990 scope.go:117] "RemoveContainer" containerID="ca4307683ea7333a625e2203775fa26ab0df3b188bf9bac1cd9b6194d3df672f" Dec 03 12:56:34 crc kubenswrapper[4990]: I1203 12:56:34.896857 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" event={"ID":"64f072f3-940c-431d-8e0f-b77e9349e79e","Type":"ContainerStarted","Data":"2c236d65302695652d7698321d2edcd44ae20c42d139eca5380d28b6567ce13b"} Dec 03 12:56:35 crc kubenswrapper[4990]: E1203 12:56:35.482627 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 12:56:35 crc kubenswrapper[4990]: E1203 12:56:35.483127 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wrsvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-x7jtx_openstack-operators(3443975f-d482-422a-a010-9940c2b1182f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:56:35 crc kubenswrapper[4990]: E1203 12:56:35.484469 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" podUID="3443975f-d482-422a-a010-9940c2b1182f" Dec 03 12:56:35 crc kubenswrapper[4990]: I1203 12:56:35.909376 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" event={"ID":"213fe9a3-0270-4ec5-b380-2f28159b5b6a","Type":"ContainerStarted","Data":"496d393e188287d29be0a9864bebbaa3c7cf3d50be49e398d039850739514c67"} Dec 03 12:56:35 crc kubenswrapper[4990]: I1203 12:56:35.911886 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" event={"ID":"425132be-e04e-46f6-ac06-1546c5b12a26","Type":"ContainerStarted","Data":"94d177ee7bbc4c9dcf7f0e7b2fede3b15a0ddc652327346de1926cd8bf1e1412"} Dec 03 12:56:35 crc kubenswrapper[4990]: I1203 12:56:35.912045 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:35 crc kubenswrapper[4990]: I1203 12:56:35.953243 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" podStartSLOduration=41.953227654 podStartE2EDuration="41.953227654s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:56:35.953000127 +0000 UTC m=+1144.094911366" watchObservedRunningTime="2025-12-03 12:56:35.953227654 +0000 UTC m=+1144.095138883" Dec 03 12:56:40 crc kubenswrapper[4990]: I1203 12:56:40.952882 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"af393e5087a271fce924a6dbdf31ad1b2b0bea5f1a2211228838f21a6a0541a1"} Dec 03 12:56:41 crc kubenswrapper[4990]: I1203 12:56:41.286210 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5b474d9948-p5wjj" Dec 03 12:56:41 crc kubenswrapper[4990]: I1203 12:56:41.960051 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" event={"ID":"0e880282-b396-4fa9-a3ac-89d81315ecd8","Type":"ContainerStarted","Data":"4b15481dc48008b9a7363eb695f0b614135bd4f3f52d4f61a4c31a62ed0989de"} Dec 03 12:56:42 crc kubenswrapper[4990]: E1203 12:56:42.944192 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" podUID="4481645b-1e70-415a-a01b-131b1761dd19" Dec 03 12:56:42 crc kubenswrapper[4990]: I1203 12:56:42.971286 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" event={"ID":"43c2b7af-9659-4ae6-8809-27475b31c611","Type":"ContainerStarted","Data":"e308475871316ffa0c2851c526de958727e49d8160a257e0c0c339f5c5fb4381"} Dec 03 12:56:42 crc kubenswrapper[4990]: I1203 12:56:42.973581 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" event={"ID":"4481645b-1e70-415a-a01b-131b1761dd19","Type":"ContainerStarted","Data":"be864ee717c83f5707d6ebbbc94fbc49a7e16c26d5078d1fa7277d21ee4c5f2e"} Dec 03 12:56:42 crc kubenswrapper[4990]: I1203 12:56:42.976727 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" event={"ID":"c0833459-1161-4beb-ad68-07d51b5b33d7","Type":"ContainerStarted","Data":"87325ec67f170209000a2f71d029431dbc05e54f842732ae82d3a12db0d7fe7c"} Dec 03 12:56:43 crc kubenswrapper[4990]: E1203 12:56:43.094107 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" podUID="4481645b-1e70-415a-a01b-131b1761dd19" Dec 03 12:56:43 crc kubenswrapper[4990]: E1203 12:56:43.725161 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" podUID="cc2d7750-b3b5-4585-95fd-4f84e1c437ad" Dec 03 12:56:44 crc kubenswrapper[4990]: E1203 12:56:44.853826 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" podUID="8bf181cd-d802-43c1-a7f0-7df5ec5008a1" Dec 03 12:56:44 crc kubenswrapper[4990]: E1203 12:56:44.890706 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" podUID="a18809dc-c194-44b4-84e8-26e6da515bb7" Dec 03 12:56:44 crc kubenswrapper[4990]: E1203 12:56:44.963476 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" podUID="d2c9d382-f60c-4c2c-a684-ddec4371b165" Dec 03 12:56:44 crc kubenswrapper[4990]: E1203 12:56:44.964598 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" podUID="6ca15749-37c2-477f-8a9a-33c07b97aaf3" Dec 03 12:56:44 crc kubenswrapper[4990]: I1203 12:56:44.991988 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" event={"ID":"0e880282-b396-4fa9-a3ac-89d81315ecd8","Type":"ContainerStarted","Data":"c7cbfef80a2200875d422541898e1d7edd04361a066cd30366d6fddf3d82e61f"} Dec 03 12:56:44 crc kubenswrapper[4990]: I1203 12:56:44.992841 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:56:44 crc kubenswrapper[4990]: I1203 12:56:44.995089 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" event={"ID":"6ca15749-37c2-477f-8a9a-33c07b97aaf3","Type":"ContainerStarted","Data":"321a4f3871f02af2d0f294a70a52ddb69a4483de4d997a9a57dc1d07c63c5e16"} Dec 03 12:56:44 crc kubenswrapper[4990]: I1203 12:56:44.997587 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" event={"ID":"c0833459-1161-4beb-ad68-07d51b5b33d7","Type":"ContainerStarted","Data":"3c5cc4c38937e1ddb948b1793326ec180eaa48a1bb17b8c741936aabb50005fd"} Dec 03 12:56:44 crc kubenswrapper[4990]: I1203 12:56:44.997763 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:56:44 crc kubenswrapper[4990]: I1203 12:56:44.999219 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" event={"ID":"8bf181cd-d802-43c1-a7f0-7df5ec5008a1","Type":"ContainerStarted","Data":"a4548f034169240635703946e5acea4cdac97bd4e51309514bc04a3cd3db0992"} Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.000792 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" event={"ID":"d2c9d382-f60c-4c2c-a684-ddec4371b165","Type":"ContainerStarted","Data":"ccb073124d70566e4288f7476011950db03113cba0071091510e0cbbdbe4a38b"} Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.002219 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" event={"ID":"93f0e562-5973-4c2a-ab48-aa167ae49ffa","Type":"ContainerStarted","Data":"83f39eca7e433016f015381cc62d4b9c326723d24dcdad65c2a3237460cddef9"} Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.002866 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.004017 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" event={"ID":"a18809dc-c194-44b4-84e8-26e6da515bb7","Type":"ContainerStarted","Data":"8d1c3c5d527ad05659fb3a43d9f6b59618de15fc40752dffa4db32e3eceb5dfe"} Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.005916 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.016491 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" podStartSLOduration=42.597682466 podStartE2EDuration="52.016441126s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:56:30.451342867 +0000 UTC m=+1138.593254096" lastFinishedPulling="2025-12-03 12:56:39.870101537 +0000 UTC m=+1148.012012756" observedRunningTime="2025-12-03 12:56:45.011942717 +0000 UTC m=+1153.153853966" watchObservedRunningTime="2025-12-03 12:56:45.016441126 +0000 UTC m=+1153.158352355" Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.069701 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" podStartSLOduration=41.246768379 podStartE2EDuration="51.069683146s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:56:30.46021984 +0000 UTC m=+1138.602131069" lastFinishedPulling="2025-12-03 12:56:40.283134607 +0000 UTC m=+1148.425045836" observedRunningTime="2025-12-03 12:56:45.063942816 +0000 UTC m=+1153.205854055" watchObservedRunningTime="2025-12-03 12:56:45.069683146 +0000 UTC m=+1153.211594375" Dec 03 12:56:45 crc kubenswrapper[4990]: I1203 12:56:45.086686 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-hqhfc" podStartSLOduration=6.058618615 podStartE2EDuration="52.086670743s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:55:55.938350745 +0000 UTC m=+1104.080261964" lastFinishedPulling="2025-12-03 12:56:41.966402853 +0000 UTC m=+1150.108314092" observedRunningTime="2025-12-03 12:56:45.08350924 +0000 UTC m=+1153.225420479" watchObservedRunningTime="2025-12-03 12:56:45.086670743 +0000 UTC m=+1153.228581972" Dec 03 12:56:45 crc kubenswrapper[4990]: E1203 12:56:45.372008 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" podUID="d2c9d382-f60c-4c2c-a684-ddec4371b165" Dec 03 12:56:45 crc kubenswrapper[4990]: E1203 12:56:45.372014 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" podUID="8bf181cd-d802-43c1-a7f0-7df5ec5008a1" Dec 03 12:56:46 crc kubenswrapper[4990]: I1203 12:56:46.013039 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" event={"ID":"e944cb60-4e1a-448f-ab25-60ff87e3a166","Type":"ContainerStarted","Data":"98e100b90df5c2ee1bb7c622e3d1d79d388733b77ed0b7058b1a18aa927d0071"} Dec 03 12:56:46 crc kubenswrapper[4990]: I1203 12:56:46.019688 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zhgnf" Dec 03 12:56:46 crc kubenswrapper[4990]: I1203 12:56:46.037796 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" podStartSLOduration=4.696167429 podStartE2EDuration="52.037774449s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.403523681 +0000 UTC m=+1104.545434910" lastFinishedPulling="2025-12-03 12:56:43.745130701 +0000 UTC m=+1151.887041930" observedRunningTime="2025-12-03 12:56:46.031524745 +0000 UTC m=+1154.173435974" watchObservedRunningTime="2025-12-03 12:56:46.037774449 +0000 UTC m=+1154.179685678" Dec 03 12:56:47 crc kubenswrapper[4990]: I1203 12:56:47.021093 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:56:47 crc kubenswrapper[4990]: I1203 12:56:47.024276 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vkrn9" Dec 03 12:56:47 crc kubenswrapper[4990]: I1203 12:56:47.030108 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9" Dec 03 12:56:48 crc kubenswrapper[4990]: I1203 12:56:48.032241 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" event={"ID":"43c2b7af-9659-4ae6-8809-27475b31c611","Type":"ContainerStarted","Data":"6662607de3bdf9967d0f74f9b19e7d5366a1c76dbc9a78afe2bfd39b19f3ba4e"} Dec 03 12:56:49 crc kubenswrapper[4990]: I1203 12:56:49.040346 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:56:49 crc kubenswrapper[4990]: I1203 12:56:49.042244 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" Dec 03 12:56:49 crc kubenswrapper[4990]: I1203 12:56:49.058185 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-256rj" podStartSLOduration=10.923500749 podStartE2EDuration="55.058153381s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.399129365 +0000 UTC m=+1104.541040594" lastFinishedPulling="2025-12-03 12:56:40.533781987 +0000 UTC m=+1148.675693226" observedRunningTime="2025-12-03 12:56:49.055249555 +0000 UTC m=+1157.197160784" watchObservedRunningTime="2025-12-03 12:56:49.058153381 +0000 UTC m=+1157.200064610" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.061002 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" event={"ID":"c7a7a3b2-d356-4dfa-b640-ae39aa2d4840","Type":"ContainerStarted","Data":"a2aea7cf036c6d25f79f18ec0e05131bcb0b6355b09b5ed69f4ca82c2a2abe32"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.061669 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.063848 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" event={"ID":"a18809dc-c194-44b4-84e8-26e6da515bb7","Type":"ContainerStarted","Data":"436e684f918cb02dd52020d03fc0805f2fa07cd7c4ca452eada6ac452c04b905"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.064024 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.068368 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" event={"ID":"96da719f-5659-45eb-ac48-37165c936527","Type":"ContainerStarted","Data":"bdfdc1616ad977f6ad6bc783610c7c7d30907e7f15bbe218382b3825092627e2"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.068652 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.071537 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.073731 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" event={"ID":"ac715eff-4f7f-44b0-b5ca-c5d283edca36","Type":"ContainerStarted","Data":"4c9c79437df02cbcddb8d1c5c28e6a665bae7fb62dbe00a2ed9982b90f68dfa3"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.073936 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.075955 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" event={"ID":"213fe9a3-0270-4ec5-b380-2f28159b5b6a","Type":"ContainerStarted","Data":"34cbcbc441e4dd1478f1c04341ff1d9d98cf1c084bd06f42de66fc9742ef3c7e"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.077547 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.077602 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.078138 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" event={"ID":"6ca15749-37c2-477f-8a9a-33c07b97aaf3","Type":"ContainerStarted","Data":"51c74cd2f15b488c846f0535157827f4f7ec4031aeb8997891355caf92af2de0"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.078905 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.079733 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.079980 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" event={"ID":"0a2be3dd-196c-4186-ac5d-85ce550aa201","Type":"ContainerStarted","Data":"a70aeaabd88f816a41de732c20c66dec86ce9510b2e003490ff8cea2212ef819"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.086305 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" event={"ID":"9106e41f-5788-46a2-86ce-6362832d6df7","Type":"ContainerStarted","Data":"9623b1a1575b62c9ea06b808599fd4abab7addeac0ad04fb107a9f9068b939ba"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.086743 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.088332 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" event={"ID":"3443975f-d482-422a-a010-9940c2b1182f","Type":"ContainerStarted","Data":"6821ac629637e8168c8e9c268fddfa13ebe05bf707448734dd9e3ce194cd6e7c"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.088379 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" event={"ID":"3443975f-d482-422a-a010-9940c2b1182f","Type":"ContainerStarted","Data":"eb8bdf565bdee656b5dec444890a3c08630928c6d6c5ca44041ec1fa20eb2d33"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.088860 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.089967 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.090574 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" event={"ID":"a8fb4bb3-4b67-4ada-8876-389bff0e9836","Type":"ContainerStarted","Data":"f4c4a64f717c34f86cb21664559cb8c120082ed038d000b09392261c7dab5a3c"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.092615 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.094351 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.094894 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" event={"ID":"64f072f3-940c-431d-8e0f-b77e9349e79e","Type":"ContainerStarted","Data":"1cfc4ada46641358310554ae44c1e82e4bf28f951c574143b20c07fc822ca870"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.096490 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.100762 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.101593 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" event={"ID":"4881a1ed-a968-4f20-a8e6-94a91ec5eceb","Type":"ContainerStarted","Data":"978716c3503620390b649396a5a8fe9347fa3a40fc6aad0235236f4b466005f7"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.103413 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.107440 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" event={"ID":"c45d3b36-d188-4bb1-b19f-5521709f572a","Type":"ContainerStarted","Data":"ae7ef371b66124843cd4b0945d969497fae58526a0516a6152a8123038dd8535"} Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.107508 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.109024 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.112160 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.127741 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-79xxb" podStartSLOduration=3.597819352 podStartE2EDuration="57.127728215s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.403405998 +0000 UTC m=+1104.545317227" lastFinishedPulling="2025-12-03 12:56:49.933314861 +0000 UTC m=+1158.075226090" observedRunningTime="2025-12-03 12:56:51.126893494 +0000 UTC m=+1159.268804723" watchObservedRunningTime="2025-12-03 12:56:51.127728215 +0000 UTC m=+1159.269639444" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.129547 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4sp89" podStartSLOduration=3.556715062 podStartE2EDuration="58.129539824s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:55:55.367408651 +0000 UTC m=+1103.509319890" lastFinishedPulling="2025-12-03 12:56:49.940233423 +0000 UTC m=+1158.082144652" observedRunningTime="2025-12-03 12:56:51.104759512 +0000 UTC m=+1159.246670741" watchObservedRunningTime="2025-12-03 12:56:51.129539824 +0000 UTC m=+1159.271451053" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.154683 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" podStartSLOduration=3.603644643 podStartE2EDuration="57.154666834s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.386182333 +0000 UTC m=+1104.528093562" lastFinishedPulling="2025-12-03 12:56:49.937204524 +0000 UTC m=+1158.079115753" observedRunningTime="2025-12-03 12:56:51.153564545 +0000 UTC m=+1159.295475774" watchObservedRunningTime="2025-12-03 12:56:51.154666834 +0000 UTC m=+1159.296578143" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.242213 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-8zwv5" podStartSLOduration=4.697523853 podStartE2EDuration="58.242196385s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.398305243 +0000 UTC m=+1104.540216472" lastFinishedPulling="2025-12-03 12:56:49.942977775 +0000 UTC m=+1158.084889004" observedRunningTime="2025-12-03 12:56:51.240844329 +0000 UTC m=+1159.382755578" watchObservedRunningTime="2025-12-03 12:56:51.242196385 +0000 UTC m=+1159.384107614" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.248915 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xt2lq" podStartSLOduration=4.702403409 podStartE2EDuration="58.248896821s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.386757488 +0000 UTC m=+1104.528668717" lastFinishedPulling="2025-12-03 12:56:49.9332509 +0000 UTC m=+1158.075162129" observedRunningTime="2025-12-03 12:56:51.219472798 +0000 UTC m=+1159.361384037" watchObservedRunningTime="2025-12-03 12:56:51.248896821 +0000 UTC m=+1159.390808070" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.286981 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" podStartSLOduration=5.242754228 podStartE2EDuration="57.286955232s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.510542072 +0000 UTC m=+1104.652453301" lastFinishedPulling="2025-12-03 12:56:48.554743076 +0000 UTC m=+1156.696654305" observedRunningTime="2025-12-03 12:56:51.275532192 +0000 UTC m=+1159.417443421" watchObservedRunningTime="2025-12-03 12:56:51.286955232 +0000 UTC m=+1159.428866461" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.313215 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" podStartSLOduration=3.858085835 podStartE2EDuration="57.313196842s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.399623838 +0000 UTC m=+1104.541535067" lastFinishedPulling="2025-12-03 12:56:49.854734845 +0000 UTC m=+1157.996646074" observedRunningTime="2025-12-03 12:56:51.309070574 +0000 UTC m=+1159.450981813" watchObservedRunningTime="2025-12-03 12:56:51.313196842 +0000 UTC m=+1159.455108081" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.373243 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4db2x" podStartSLOduration=5.314230917 podStartE2EDuration="57.3732186s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.495753103 +0000 UTC m=+1104.637664332" lastFinishedPulling="2025-12-03 12:56:48.554740776 +0000 UTC m=+1156.696652015" observedRunningTime="2025-12-03 12:56:51.344830294 +0000 UTC m=+1159.486741523" watchObservedRunningTime="2025-12-03 12:56:51.3732186 +0000 UTC m=+1159.515129839" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.374889 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qt45w" podStartSLOduration=3.823774012 podStartE2EDuration="57.374880164s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.386536903 +0000 UTC m=+1104.528448212" lastFinishedPulling="2025-12-03 12:56:49.937643135 +0000 UTC m=+1158.079554364" observedRunningTime="2025-12-03 12:56:51.372709486 +0000 UTC m=+1159.514620715" watchObservedRunningTime="2025-12-03 12:56:51.374880164 +0000 UTC m=+1159.516791393" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.451404 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-z484w" podStartSLOduration=4.368318198 podStartE2EDuration="58.451376515s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:55:55.879183135 +0000 UTC m=+1104.021094374" lastFinishedPulling="2025-12-03 12:56:49.962241462 +0000 UTC m=+1158.104152691" observedRunningTime="2025-12-03 12:56:51.413739256 +0000 UTC m=+1159.555650485" watchObservedRunningTime="2025-12-03 12:56:51.451376515 +0000 UTC m=+1159.593287754" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.455433 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-4wkxw" podStartSLOduration=5.878714813 podStartE2EDuration="57.455411821s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.574712014 +0000 UTC m=+1104.716623243" lastFinishedPulling="2025-12-03 12:56:48.151409022 +0000 UTC m=+1156.293320251" observedRunningTime="2025-12-03 12:56:51.449060874 +0000 UTC m=+1159.590972103" watchObservedRunningTime="2025-12-03 12:56:51.455411821 +0000 UTC m=+1159.597323050" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.485928 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" podStartSLOduration=4.21684889 podStartE2EDuration="57.485907333s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.584708168 +0000 UTC m=+1104.726619397" lastFinishedPulling="2025-12-03 12:56:49.853766611 +0000 UTC m=+1157.995677840" observedRunningTime="2025-12-03 12:56:51.475841218 +0000 UTC m=+1159.617752457" watchObservedRunningTime="2025-12-03 12:56:51.485907333 +0000 UTC m=+1159.627818562" Dec 03 12:56:51 crc kubenswrapper[4990]: I1203 12:56:51.508704 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7n65k" podStartSLOduration=3.936995449 podStartE2EDuration="58.508686032s" podCreationTimestamp="2025-12-03 12:55:53 +0000 UTC" firstStartedPulling="2025-12-03 12:55:55.367410691 +0000 UTC m=+1103.509321920" lastFinishedPulling="2025-12-03 12:56:49.939101284 +0000 UTC m=+1158.081012503" observedRunningTime="2025-12-03 12:56:51.505821166 +0000 UTC m=+1159.647732415" watchObservedRunningTime="2025-12-03 12:56:51.508686032 +0000 UTC m=+1159.650597261" Dec 03 12:56:52 crc kubenswrapper[4990]: I1203 12:56:52.118740 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:56:52 crc kubenswrapper[4990]: I1203 12:56:52.119126 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:56:52 crc kubenswrapper[4990]: I1203 12:56:52.119204 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-c8cgl" Dec 03 12:56:55 crc kubenswrapper[4990]: I1203 12:56:55.088976 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xg94h" Dec 03 12:56:55 crc kubenswrapper[4990]: I1203 12:56:55.272494 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dg8m2" Dec 03 12:56:55 crc kubenswrapper[4990]: I1203 12:56:55.486229 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-x7jtx" Dec 03 12:56:57 crc kubenswrapper[4990]: I1203 12:56:57.164068 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" event={"ID":"cc2d7750-b3b5-4585-95fd-4f84e1c437ad","Type":"ContainerStarted","Data":"674ccc62cb2874f85c619196ade326599c22579c0730fe22bc3a658e22b2b83b"} Dec 03 12:56:57 crc kubenswrapper[4990]: I1203 12:56:57.185904 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-p8hpt" podStartSLOduration=3.795545516 podStartE2EDuration="1m3.185876348s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.609160113 +0000 UTC m=+1104.751071342" lastFinishedPulling="2025-12-03 12:56:55.999490945 +0000 UTC m=+1164.141402174" observedRunningTime="2025-12-03 12:56:57.181903553 +0000 UTC m=+1165.323814782" watchObservedRunningTime="2025-12-03 12:56:57.185876348 +0000 UTC m=+1165.327787577" Dec 03 12:56:59 crc kubenswrapper[4990]: I1203 12:56:59.185934 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" event={"ID":"4481645b-1e70-415a-a01b-131b1761dd19","Type":"ContainerStarted","Data":"97b27a9e779153d87ca46dbee37d06a5a3e32b1db5d2061c9d5b440580cb2494"} Dec 03 12:56:59 crc kubenswrapper[4990]: I1203 12:56:59.187149 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:56:59 crc kubenswrapper[4990]: I1203 12:56:59.189318 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" event={"ID":"d2c9d382-f60c-4c2c-a684-ddec4371b165","Type":"ContainerStarted","Data":"6fa52c14647ad55d80d9f34c33e3b7d530894910a18bb9350c3124b874fb92d0"} Dec 03 12:56:59 crc kubenswrapper[4990]: I1203 12:56:59.189548 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:56:59 crc kubenswrapper[4990]: I1203 12:56:59.207435 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" podStartSLOduration=3.406465025 podStartE2EDuration="1m5.207420009s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.591106707 +0000 UTC m=+1104.733017936" lastFinishedPulling="2025-12-03 12:56:58.392061691 +0000 UTC m=+1166.533972920" observedRunningTime="2025-12-03 12:56:59.204704367 +0000 UTC m=+1167.346615596" watchObservedRunningTime="2025-12-03 12:56:59.207420009 +0000 UTC m=+1167.349331238" Dec 03 12:56:59 crc kubenswrapper[4990]: I1203 12:56:59.228141 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" podStartSLOduration=3.333531967 podStartE2EDuration="1m5.228122333s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.601621603 +0000 UTC m=+1104.743532832" lastFinishedPulling="2025-12-03 12:56:58.496211969 +0000 UTC m=+1166.638123198" observedRunningTime="2025-12-03 12:56:59.226621404 +0000 UTC m=+1167.368532633" watchObservedRunningTime="2025-12-03 12:56:59.228122333 +0000 UTC m=+1167.370033562" Dec 03 12:57:01 crc kubenswrapper[4990]: I1203 12:57:01.207803 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" event={"ID":"8bf181cd-d802-43c1-a7f0-7df5ec5008a1","Type":"ContainerStarted","Data":"9eead4261a8dfa919082c7ffbde6cb2fa1a705ef29e3159352c2b6058c65eab4"} Dec 03 12:57:01 crc kubenswrapper[4990]: I1203 12:57:01.208812 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:57:01 crc kubenswrapper[4990]: I1203 12:57:01.236889 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" podStartSLOduration=3.289353445 podStartE2EDuration="1m7.236843596s" podCreationTimestamp="2025-12-03 12:55:54 +0000 UTC" firstStartedPulling="2025-12-03 12:55:56.594412443 +0000 UTC m=+1104.736323672" lastFinishedPulling="2025-12-03 12:57:00.541902594 +0000 UTC m=+1168.683813823" observedRunningTime="2025-12-03 12:57:01.228873436 +0000 UTC m=+1169.370784685" watchObservedRunningTime="2025-12-03 12:57:01.236843596 +0000 UTC m=+1169.378754825" Dec 03 12:57:05 crc kubenswrapper[4990]: I1203 12:57:05.550137 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-rvslq" Dec 03 12:57:05 crc kubenswrapper[4990]: I1203 12:57:05.594850 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7cl8r" Dec 03 12:57:05 crc kubenswrapper[4990]: I1203 12:57:05.619598 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tg4ww" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.715906 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2lj9"] Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.718035 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.729072 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.729389 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-57tv4" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.729607 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.729679 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.734295 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2lj9"] Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.778098 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zzwqx"] Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.783031 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.786299 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.794212 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zzwqx"] Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.831877 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn9pt\" (UniqueName: \"kubernetes.io/projected/523066ca-dde9-4160-b20f-d7e05a8f564d-kube-api-access-kn9pt\") pod \"dnsmasq-dns-675f4bcbfc-s2lj9\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.831965 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523066ca-dde9-4160-b20f-d7e05a8f564d-config\") pod \"dnsmasq-dns-675f4bcbfc-s2lj9\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.932771 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn9pt\" (UniqueName: \"kubernetes.io/projected/523066ca-dde9-4160-b20f-d7e05a8f564d-kube-api-access-kn9pt\") pod \"dnsmasq-dns-675f4bcbfc-s2lj9\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.932867 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.932898 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523066ca-dde9-4160-b20f-d7e05a8f564d-config\") pod \"dnsmasq-dns-675f4bcbfc-s2lj9\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.932947 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbvn\" (UniqueName: \"kubernetes.io/projected/503c78af-26fd-4428-b885-8398b6343624-kube-api-access-kqbvn\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.932970 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-config\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.934033 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523066ca-dde9-4160-b20f-d7e05a8f564d-config\") pod \"dnsmasq-dns-675f4bcbfc-s2lj9\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:21 crc kubenswrapper[4990]: I1203 12:57:21.953684 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn9pt\" (UniqueName: \"kubernetes.io/projected/523066ca-dde9-4160-b20f-d7e05a8f564d-kube-api-access-kn9pt\") pod \"dnsmasq-dns-675f4bcbfc-s2lj9\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.034589 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.034725 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbvn\" (UniqueName: \"kubernetes.io/projected/503c78af-26fd-4428-b885-8398b6343624-kube-api-access-kqbvn\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.034747 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-config\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.035578 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.035842 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-config\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.044040 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.051885 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbvn\" (UniqueName: \"kubernetes.io/projected/503c78af-26fd-4428-b885-8398b6343624-kube-api-access-kqbvn\") pod \"dnsmasq-dns-78dd6ddcc-zzwqx\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.118026 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.532361 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2lj9"] Dec 03 12:57:22 crc kubenswrapper[4990]: I1203 12:57:22.628255 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zzwqx"] Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.367169 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" event={"ID":"503c78af-26fd-4428-b885-8398b6343624","Type":"ContainerStarted","Data":"f6c4105879d51148e0630fb6315a735d178966ab0c39c57dfdb3bb0087b20851"} Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.368502 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" event={"ID":"523066ca-dde9-4160-b20f-d7e05a8f564d","Type":"ContainerStarted","Data":"88aaedfd0e8f291f6711e29996a55e816a050dcff816cd774cf3c43f32339b3a"} Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.812634 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2lj9"] Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.864660 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-c974j"] Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.896599 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.912633 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-c974j"] Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.970461 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-dns-svc\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.970603 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt525\" (UniqueName: \"kubernetes.io/projected/4b643524-7647-40ed-9151-90c0921700cf-kube-api-access-nt525\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:23 crc kubenswrapper[4990]: I1203 12:57:23.970741 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-config\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.072283 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-config\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.072391 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-dns-svc\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.072443 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt525\" (UniqueName: \"kubernetes.io/projected/4b643524-7647-40ed-9151-90c0921700cf-kube-api-access-nt525\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.073332 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-config\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.073791 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-dns-svc\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.148147 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt525\" (UniqueName: \"kubernetes.io/projected/4b643524-7647-40ed-9151-90c0921700cf-kube-api-access-nt525\") pod \"dnsmasq-dns-666b6646f7-c974j\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.210101 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zzwqx"] Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.228019 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.241026 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dh6zz"] Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.242737 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.267077 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dh6zz"] Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.379387 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.379909 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn2zc\" (UniqueName: \"kubernetes.io/projected/5c617111-b3c2-431b-80eb-c3c30f0b4269-kube-api-access-qn2zc\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.380080 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-config\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.481563 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.481618 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn2zc\" (UniqueName: \"kubernetes.io/projected/5c617111-b3c2-431b-80eb-c3c30f0b4269-kube-api-access-qn2zc\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.482815 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.483148 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-config\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.484039 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-config\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.511826 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn2zc\" (UniqueName: \"kubernetes.io/projected/5c617111-b3c2-431b-80eb-c3c30f0b4269-kube-api-access-qn2zc\") pod \"dnsmasq-dns-57d769cc4f-dh6zz\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.589477 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:57:24 crc kubenswrapper[4990]: I1203 12:57:24.831834 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-c974j"] Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.046236 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.053323 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.055470 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.055834 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mtzzs" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.058853 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.062333 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.062385 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.062437 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.062660 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.062895 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.124681 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dh6zz"] Dec 03 12:57:25 crc kubenswrapper[4990]: W1203 12:57:25.135283 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c617111_b3c2_431b_80eb_c3c30f0b4269.slice/crio-6d22ce3beccef60ff39435f3a45565d6ca21ceab6952aa105d38be28befb7677 WatchSource:0}: Error finding container 6d22ce3beccef60ff39435f3a45565d6ca21ceab6952aa105d38be28befb7677: Status 404 returned error can't find the container with id 6d22ce3beccef60ff39435f3a45565d6ca21ceab6952aa105d38be28befb7677 Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197546 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197604 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c414845-915f-47b6-8a52-0e6f4e9ad886-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197628 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-config-data\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197644 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197664 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197686 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c414845-915f-47b6-8a52-0e6f4e9ad886-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197712 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197873 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.197974 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf6sm\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-kube-api-access-bf6sm\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.198025 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.198122 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299788 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299845 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-config-data\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299876 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299902 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c414845-915f-47b6-8a52-0e6f4e9ad886-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299932 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299957 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.299988 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf6sm\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-kube-api-access-bf6sm\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.300013 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.300051 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.300091 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.300110 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c414845-915f-47b6-8a52-0e6f4e9ad886-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.302241 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-config-data\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.302560 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.302835 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.303061 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.303755 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.304137 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.310694 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c414845-915f-47b6-8a52-0e6f4e9ad886-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.312125 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c414845-915f-47b6-8a52-0e6f4e9ad886-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.312296 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.326748 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.329751 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf6sm\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-kube-api-access-bf6sm\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.345302 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.385165 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.392147 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.399859 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.405141 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.408182 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.408228 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-j442g" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.408617 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.408647 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.408717 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.408952 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.412303 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" event={"ID":"5c617111-b3c2-431b-80eb-c3c30f0b4269","Type":"ContainerStarted","Data":"6d22ce3beccef60ff39435f3a45565d6ca21ceab6952aa105d38be28befb7677"} Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.413163 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.418283 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-c974j" event={"ID":"4b643524-7647-40ed-9151-90c0921700cf","Type":"ContainerStarted","Data":"fa2cc7f824400945b997133b70aa02f0c00943ff7d7322b4fe10303844885144"} Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.503701 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d0a6648-1fc6-4928-ab9f-19829eaa0726-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.503779 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.503846 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnjz5\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-kube-api-access-xnjz5\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.503892 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.503931 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.503959 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.504005 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.504040 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.504095 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d0a6648-1fc6-4928-ab9f-19829eaa0726-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.504126 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.504175 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608338 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608795 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608834 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d0a6648-1fc6-4928-ab9f-19829eaa0726-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608860 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608885 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnjz5\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-kube-api-access-xnjz5\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608914 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608942 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.608988 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.609320 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.609367 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.609420 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.609478 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.609500 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.609791 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.610221 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.611521 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.611555 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d0a6648-1fc6-4928-ab9f-19829eaa0726-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.616100 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d0a6648-1fc6-4928-ab9f-19829eaa0726-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.616530 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.620095 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d0a6648-1fc6-4928-ab9f-19829eaa0726-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.620149 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.637204 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnjz5\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-kube-api-access-xnjz5\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.649964 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.748293 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:57:25 crc kubenswrapper[4990]: I1203 12:57:25.961217 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 12:57:25 crc kubenswrapper[4990]: W1203 12:57:25.968546 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c414845_915f_47b6_8a52_0e6f4e9ad886.slice/crio-c54fe6cf6c4228759e0ea76c010664c1df5fc18423828e31772c0fc231e1d5b8 WatchSource:0}: Error finding container c54fe6cf6c4228759e0ea76c010664c1df5fc18423828e31772c0fc231e1d5b8: Status 404 returned error can't find the container with id c54fe6cf6c4228759e0ea76c010664c1df5fc18423828e31772c0fc231e1d5b8 Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.298837 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.441104 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d0a6648-1fc6-4928-ab9f-19829eaa0726","Type":"ContainerStarted","Data":"4b38f92942fcc00996e1422c0650b0fe008bd05143650cb6c15aca4ce203b726"} Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.444420 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5c414845-915f-47b6-8a52-0e6f4e9ad886","Type":"ContainerStarted","Data":"c54fe6cf6c4228759e0ea76c010664c1df5fc18423828e31772c0fc231e1d5b8"} Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.807153 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.809398 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.813312 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-pv99c" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.813603 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.813862 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.821262 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.840875 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.844807 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943380 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fb6a200-beee-4b47-ac8d-e370e307070f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943464 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-kolla-config\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943515 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fb6a200-beee-4b47-ac8d-e370e307070f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943574 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-config-data-default\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943833 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4fb6a200-beee-4b47-ac8d-e370e307070f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943899 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd6ln\" (UniqueName: \"kubernetes.io/projected/4fb6a200-beee-4b47-ac8d-e370e307070f-kube-api-access-nd6ln\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943953 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:26 crc kubenswrapper[4990]: I1203 12:57:26.943989 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.050831 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-config-data-default\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051014 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4fb6a200-beee-4b47-ac8d-e370e307070f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051048 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd6ln\" (UniqueName: \"kubernetes.io/projected/4fb6a200-beee-4b47-ac8d-e370e307070f-kube-api-access-nd6ln\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051114 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051145 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051250 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fb6a200-beee-4b47-ac8d-e370e307070f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051309 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-kolla-config\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.051327 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fb6a200-beee-4b47-ac8d-e370e307070f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.053428 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.054112 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4fb6a200-beee-4b47-ac8d-e370e307070f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.054643 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-kolla-config\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.054970 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-config-data-default\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.055674 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fb6a200-beee-4b47-ac8d-e370e307070f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.061125 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fb6a200-beee-4b47-ac8d-e370e307070f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.062414 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fb6a200-beee-4b47-ac8d-e370e307070f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.079392 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd6ln\" (UniqueName: \"kubernetes.io/projected/4fb6a200-beee-4b47-ac8d-e370e307070f-kube-api-access-nd6ln\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.085224 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"4fb6a200-beee-4b47-ac8d-e370e307070f\") " pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.173351 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 12:57:27 crc kubenswrapper[4990]: I1203 12:57:27.847523 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.448668 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.450380 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.454646 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.454724 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.455256 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-ng8d8" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.455887 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.466070 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.499888 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4fb6a200-beee-4b47-ac8d-e370e307070f","Type":"ContainerStarted","Data":"1aff1a16032b4aff00719e9d938cdbf4820b3e8b305e1a76586130aa6cf1edb8"} Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593543 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593600 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgdd9\" (UniqueName: \"kubernetes.io/projected/81e14867-9f93-4820-ba0f-c6dbd12e31db-kube-api-access-fgdd9\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593630 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593677 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/81e14867-9f93-4820-ba0f-c6dbd12e31db-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593708 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593746 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593786 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81e14867-9f93-4820-ba0f-c6dbd12e31db-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.593811 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/81e14867-9f93-4820-ba0f-c6dbd12e31db-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.598303 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.599523 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.605336 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-qgxlb" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.605606 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.606182 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.615974 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694622 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgdd9\" (UniqueName: \"kubernetes.io/projected/81e14867-9f93-4820-ba0f-c6dbd12e31db-kube-api-access-fgdd9\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694665 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dbgp\" (UniqueName: \"kubernetes.io/projected/24350652-92e4-40b6-82d5-4b785a2a0314-kube-api-access-9dbgp\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694700 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694752 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/24350652-92e4-40b6-82d5-4b785a2a0314-memcached-tls-certs\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694779 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/81e14867-9f93-4820-ba0f-c6dbd12e31db-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694809 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694828 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24350652-92e4-40b6-82d5-4b785a2a0314-config-data\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694851 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/24350652-92e4-40b6-82d5-4b785a2a0314-kolla-config\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694879 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694914 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81e14867-9f93-4820-ba0f-c6dbd12e31db-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694934 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/81e14867-9f93-4820-ba0f-c6dbd12e31db-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694972 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24350652-92e4-40b6-82d5-4b785a2a0314-combined-ca-bundle\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.694991 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.695493 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.695944 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.696413 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/81e14867-9f93-4820-ba0f-c6dbd12e31db-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.697377 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.699821 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e14867-9f93-4820-ba0f-c6dbd12e31db-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.710303 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/81e14867-9f93-4820-ba0f-c6dbd12e31db-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.712076 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81e14867-9f93-4820-ba0f-c6dbd12e31db-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.744617 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgdd9\" (UniqueName: \"kubernetes.io/projected/81e14867-9f93-4820-ba0f-c6dbd12e31db-kube-api-access-fgdd9\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.797985 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24350652-92e4-40b6-82d5-4b785a2a0314-config-data\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.798058 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/24350652-92e4-40b6-82d5-4b785a2a0314-kolla-config\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.798204 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24350652-92e4-40b6-82d5-4b785a2a0314-combined-ca-bundle\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.798234 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dbgp\" (UniqueName: \"kubernetes.io/projected/24350652-92e4-40b6-82d5-4b785a2a0314-kube-api-access-9dbgp\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.798304 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/24350652-92e4-40b6-82d5-4b785a2a0314-memcached-tls-certs\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.799918 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24350652-92e4-40b6-82d5-4b785a2a0314-config-data\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.802924 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81e14867-9f93-4820-ba0f-c6dbd12e31db\") " pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.814300 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/24350652-92e4-40b6-82d5-4b785a2a0314-kolla-config\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.827002 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/24350652-92e4-40b6-82d5-4b785a2a0314-memcached-tls-certs\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.828667 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dbgp\" (UniqueName: \"kubernetes.io/projected/24350652-92e4-40b6-82d5-4b785a2a0314-kube-api-access-9dbgp\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.829117 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24350652-92e4-40b6-82d5-4b785a2a0314-combined-ca-bundle\") pod \"memcached-0\" (UID: \"24350652-92e4-40b6-82d5-4b785a2a0314\") " pod="openstack/memcached-0" Dec 03 12:57:28 crc kubenswrapper[4990]: I1203 12:57:28.942588 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 12:57:29 crc kubenswrapper[4990]: I1203 12:57:29.098607 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.076898 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.079304 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.081824 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6dd8c" Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.106782 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.159589 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnxn9\" (UniqueName: \"kubernetes.io/projected/9896307b-5428-4f09-a960-1001d654e2be-kube-api-access-pnxn9\") pod \"kube-state-metrics-0\" (UID: \"9896307b-5428-4f09-a960-1001d654e2be\") " pod="openstack/kube-state-metrics-0" Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.260753 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnxn9\" (UniqueName: \"kubernetes.io/projected/9896307b-5428-4f09-a960-1001d654e2be-kube-api-access-pnxn9\") pod \"kube-state-metrics-0\" (UID: \"9896307b-5428-4f09-a960-1001d654e2be\") " pod="openstack/kube-state-metrics-0" Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.285792 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnxn9\" (UniqueName: \"kubernetes.io/projected/9896307b-5428-4f09-a960-1001d654e2be-kube-api-access-pnxn9\") pod \"kube-state-metrics-0\" (UID: \"9896307b-5428-4f09-a960-1001d654e2be\") " pod="openstack/kube-state-metrics-0" Dec 03 12:57:31 crc kubenswrapper[4990]: I1203 12:57:31.403443 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.824525 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6frjn"] Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.826445 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.828630 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-wgjtl" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.829091 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.829288 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.853052 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6frjn"] Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.866130 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-lbt2k"] Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.868001 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.905790 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-ovn-controller-tls-certs\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.905843 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-log\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.905869 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-run-ovn\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.905891 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-log-ovn\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.905909 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-scripts\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.905955 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-run\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906002 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f81fd15-3954-40f5-9033-18cb8e9b01ed-scripts\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906025 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-run\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906052 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-etc-ovs\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906078 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-combined-ca-bundle\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906095 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g4np\" (UniqueName: \"kubernetes.io/projected/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-kube-api-access-2g4np\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906352 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-lib\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.906442 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l687\" (UniqueName: \"kubernetes.io/projected/3f81fd15-3954-40f5-9033-18cb8e9b01ed-kube-api-access-2l687\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:33 crc kubenswrapper[4990]: I1203 12:57:33.907124 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lbt2k"] Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007599 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-ovn-controller-tls-certs\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007677 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-log\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007718 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-run-ovn\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007758 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-log-ovn\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007791 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-scripts\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007845 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-run\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007912 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f81fd15-3954-40f5-9033-18cb8e9b01ed-scripts\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007944 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-run\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.007979 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-etc-ovs\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.008017 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-combined-ca-bundle\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.008047 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g4np\" (UniqueName: \"kubernetes.io/projected/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-kube-api-access-2g4np\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.008124 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-lib\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.008173 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l687\" (UniqueName: \"kubernetes.io/projected/3f81fd15-3954-40f5-9033-18cb8e9b01ed-kube-api-access-2l687\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.009756 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-log\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.009797 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-run\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.009878 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-log-ovn\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.009892 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-run-ovn\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.010348 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-var-lib\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.010397 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-var-run\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.010564 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3f81fd15-3954-40f5-9033-18cb8e9b01ed-etc-ovs\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.012362 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-scripts\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.012713 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f81fd15-3954-40f5-9033-18cb8e9b01ed-scripts\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.015708 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-combined-ca-bundle\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.018158 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-ovn-controller-tls-certs\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.026188 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l687\" (UniqueName: \"kubernetes.io/projected/3f81fd15-3954-40f5-9033-18cb8e9b01ed-kube-api-access-2l687\") pod \"ovn-controller-ovs-lbt2k\" (UID: \"3f81fd15-3954-40f5-9033-18cb8e9b01ed\") " pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.027245 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g4np\" (UniqueName: \"kubernetes.io/projected/83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5-kube-api-access-2g4np\") pod \"ovn-controller-6frjn\" (UID: \"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5\") " pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.146022 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.196060 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.725415 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.727735 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.748238 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.749006 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.749915 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.751478 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.751587 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.751885 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k4k8b" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821184 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821276 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8ae85bd2-714b-4a62-b287-c2a9780753bd-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821314 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821347 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821389 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn5wv\" (UniqueName: \"kubernetes.io/projected/8ae85bd2-714b-4a62-b287-c2a9780753bd-kube-api-access-qn5wv\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821422 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae85bd2-714b-4a62-b287-c2a9780753bd-config\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821471 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ae85bd2-714b-4a62-b287-c2a9780753bd-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.821496 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923213 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8ae85bd2-714b-4a62-b287-c2a9780753bd-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923307 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923371 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923530 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn5wv\" (UniqueName: \"kubernetes.io/projected/8ae85bd2-714b-4a62-b287-c2a9780753bd-kube-api-access-qn5wv\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923587 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae85bd2-714b-4a62-b287-c2a9780753bd-config\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923650 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ae85bd2-714b-4a62-b287-c2a9780753bd-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923687 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923768 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.923993 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8ae85bd2-714b-4a62-b287-c2a9780753bd-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.924308 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.924594 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae85bd2-714b-4a62-b287-c2a9780753bd-config\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.925061 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ae85bd2-714b-4a62-b287-c2a9780753bd-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.930424 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.930931 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.944547 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae85bd2-714b-4a62-b287-c2a9780753bd-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.947783 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn5wv\" (UniqueName: \"kubernetes.io/projected/8ae85bd2-714b-4a62-b287-c2a9780753bd-kube-api-access-qn5wv\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:34 crc kubenswrapper[4990]: I1203 12:57:34.964970 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8ae85bd2-714b-4a62-b287-c2a9780753bd\") " pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:35 crc kubenswrapper[4990]: I1203 12:57:35.075794 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.185021 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.187091 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.191836 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.200901 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.201335 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.201614 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-5bf7v" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.201765 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375363 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375434 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63ff1335-eac5-4524-b448-785488679bff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375797 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmx6r\" (UniqueName: \"kubernetes.io/projected/63ff1335-eac5-4524-b448-785488679bff-kube-api-access-fmx6r\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375831 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63ff1335-eac5-4524-b448-785488679bff-config\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375853 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375923 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.375944 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.376440 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63ff1335-eac5-4524-b448-785488679bff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.478840 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63ff1335-eac5-4524-b448-785488679bff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.478923 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.478959 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63ff1335-eac5-4524-b448-785488679bff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.478990 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmx6r\" (UniqueName: \"kubernetes.io/projected/63ff1335-eac5-4524-b448-785488679bff-kube-api-access-fmx6r\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.479008 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63ff1335-eac5-4524-b448-785488679bff-config\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.479028 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.479067 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.479085 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.479707 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.585489 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63ff1335-eac5-4524-b448-785488679bff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.587107 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63ff1335-eac5-4524-b448-785488679bff-config\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.587807 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63ff1335-eac5-4524-b448-785488679bff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.591393 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.591451 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmx6r\" (UniqueName: \"kubernetes.io/projected/63ff1335-eac5-4524-b448-785488679bff-kube-api-access-fmx6r\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.591479 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.601630 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ff1335-eac5-4524-b448-785488679bff-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.605833 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"63ff1335-eac5-4524-b448-785488679bff\") " pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:38 crc kubenswrapper[4990]: I1203 12:57:38.880126 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 12:57:40 crc kubenswrapper[4990]: E1203 12:57:40.866721 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 12:57:40 crc kubenswrapper[4990]: E1203 12:57:40.867732 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bf6sm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(5c414845-915f-47b6-8a52-0e6f4e9ad886): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:57:40 crc kubenswrapper[4990]: E1203 12:57:40.868985 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" Dec 03 12:57:41 crc kubenswrapper[4990]: E1203 12:57:41.635027 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" Dec 03 12:58:24 crc kubenswrapper[4990]: E1203 12:58:24.898642 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 03 12:58:24 crc kubenswrapper[4990]: E1203 12:58:24.899278 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nd6ln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(4fb6a200-beee-4b47-ac8d-e370e307070f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:58:24 crc kubenswrapper[4990]: E1203 12:58:24.900559 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="4fb6a200-beee-4b47-ac8d-e370e307070f" Dec 03 12:58:25 crc kubenswrapper[4990]: E1203 12:58:25.326145 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="4fb6a200-beee-4b47-ac8d-e370e307070f" Dec 03 12:58:25 crc kubenswrapper[4990]: I1203 12:58:25.624568 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6frjn"] Dec 03 12:58:25 crc kubenswrapper[4990]: I1203 12:58:25.764123 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.298067 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.298541 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nt525,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-c974j_openstack(4b643524-7647-40ed-9151-90c0921700cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.300152 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-c974j" podUID="4b643524-7647-40ed-9151-90c0921700cf" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.334441 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.334652 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qn2zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-dh6zz_openstack(5c617111-b3c2-431b-80eb-c3c30f0b4269): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.337255 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.362311 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.362835 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kqbvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zzwqx_openstack(503c78af-26fd-4428-b885-8398b6343624): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.364971 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" podUID="503c78af-26fd-4428-b885-8398b6343624" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.401011 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.401187 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kn9pt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-s2lj9_openstack(523066ca-dde9-4160-b20f-d7e05a8f564d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:58:26 crc kubenswrapper[4990]: E1203 12:58:26.402866 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" podUID="523066ca-dde9-4160-b20f-d7e05a8f564d" Dec 03 12:58:26 crc kubenswrapper[4990]: W1203 12:58:26.727683 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9896307b_5428_4f09_a960_1001d654e2be.slice/crio-5718107ed91601b263fe6a6e9f2d8713c15e4eb710477f9164298f5371602d98 WatchSource:0}: Error finding container 5718107ed91601b263fe6a6e9f2d8713c15e4eb710477f9164298f5371602d98: Status 404 returned error can't find the container with id 5718107ed91601b263fe6a6e9f2d8713c15e4eb710477f9164298f5371602d98 Dec 03 12:58:26 crc kubenswrapper[4990]: I1203 12:58:26.729521 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 12:58:26 crc kubenswrapper[4990]: I1203 12:58:26.816952 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 12:58:26 crc kubenswrapper[4990]: W1203 12:58:26.828604 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24350652_92e4_40b6_82d5_4b785a2a0314.slice/crio-dcc39110cf2012759b1dc4af3cc7bbf68eb5578bd9c9d2c46eeaee083acbee46 WatchSource:0}: Error finding container dcc39110cf2012759b1dc4af3cc7bbf68eb5578bd9c9d2c46eeaee083acbee46: Status 404 returned error can't find the container with id dcc39110cf2012759b1dc4af3cc7bbf68eb5578bd9c9d2c46eeaee083acbee46 Dec 03 12:58:26 crc kubenswrapper[4990]: I1203 12:58:26.915281 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 12:58:26 crc kubenswrapper[4990]: W1203 12:58:26.917266 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81e14867_9f93_4820_ba0f_c6dbd12e31db.slice/crio-f7f53b99103caa9509f238820b3eaa80e5e0a9b5882313863d7220379ea01f26 WatchSource:0}: Error finding container f7f53b99103caa9509f238820b3eaa80e5e0a9b5882313863d7220379ea01f26: Status 404 returned error can't find the container with id f7f53b99103caa9509f238820b3eaa80e5e0a9b5882313863d7220379ea01f26 Dec 03 12:58:26 crc kubenswrapper[4990]: W1203 12:58:26.919758 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ae85bd2_714b_4a62_b287_c2a9780753bd.slice/crio-f83db6b0d2de0df27ad51d3eefaf1c80e1821188b0309226643a715719d09195 WatchSource:0}: Error finding container f83db6b0d2de0df27ad51d3eefaf1c80e1821188b0309226643a715719d09195: Status 404 returned error can't find the container with id f83db6b0d2de0df27ad51d3eefaf1c80e1821188b0309226643a715719d09195 Dec 03 12:58:26 crc kubenswrapper[4990]: I1203 12:58:26.925968 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.022118 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn" event={"ID":"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5","Type":"ContainerStarted","Data":"575019ed9522712e239555c9bc481407b6407d8aabd93a88de9c663b716fd414"} Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.024237 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"63ff1335-eac5-4524-b448-785488679bff","Type":"ContainerStarted","Data":"392953e09d92f469afa25114241ac5698d5d7e8dec693bb216f060bb91b65514"} Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.026125 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8ae85bd2-714b-4a62-b287-c2a9780753bd","Type":"ContainerStarted","Data":"f83db6b0d2de0df27ad51d3eefaf1c80e1821188b0309226643a715719d09195"} Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.028208 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81e14867-9f93-4820-ba0f-c6dbd12e31db","Type":"ContainerStarted","Data":"f7f53b99103caa9509f238820b3eaa80e5e0a9b5882313863d7220379ea01f26"} Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.029915 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9896307b-5428-4f09-a960-1001d654e2be","Type":"ContainerStarted","Data":"5718107ed91601b263fe6a6e9f2d8713c15e4eb710477f9164298f5371602d98"} Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.031725 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"24350652-92e4-40b6-82d5-4b785a2a0314","Type":"ContainerStarted","Data":"dcc39110cf2012759b1dc4af3cc7bbf68eb5578bd9c9d2c46eeaee083acbee46"} Dec 03 12:58:27 crc kubenswrapper[4990]: E1203 12:58:27.033642 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-c974j" podUID="4b643524-7647-40ed-9151-90c0921700cf" Dec 03 12:58:27 crc kubenswrapper[4990]: E1203 12:58:27.035797 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.048758 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lbt2k"] Dec 03 12:58:27 crc kubenswrapper[4990]: W1203 12:58:27.049242 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f81fd15_3954_40f5_9033_18cb8e9b01ed.slice/crio-b7a32782e41e59915b60a00fb31c3f8d682e0e76ccae43dec041744671ccd654 WatchSource:0}: Error finding container b7a32782e41e59915b60a00fb31c3f8d682e0e76ccae43dec041744671ccd654: Status 404 returned error can't find the container with id b7a32782e41e59915b60a00fb31c3f8d682e0e76ccae43dec041744671ccd654 Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.568224 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.575904 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.740637 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-config\") pod \"503c78af-26fd-4428-b885-8398b6343624\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.741082 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-config" (OuterVolumeSpecName: "config") pod "503c78af-26fd-4428-b885-8398b6343624" (UID: "503c78af-26fd-4428-b885-8398b6343624"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.741358 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523066ca-dde9-4160-b20f-d7e05a8f564d-config\") pod \"523066ca-dde9-4160-b20f-d7e05a8f564d\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.741500 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn9pt\" (UniqueName: \"kubernetes.io/projected/523066ca-dde9-4160-b20f-d7e05a8f564d-kube-api-access-kn9pt\") pod \"523066ca-dde9-4160-b20f-d7e05a8f564d\" (UID: \"523066ca-dde9-4160-b20f-d7e05a8f564d\") " Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.741598 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqbvn\" (UniqueName: \"kubernetes.io/projected/503c78af-26fd-4428-b885-8398b6343624-kube-api-access-kqbvn\") pod \"503c78af-26fd-4428-b885-8398b6343624\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.741692 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-dns-svc\") pod \"503c78af-26fd-4428-b885-8398b6343624\" (UID: \"503c78af-26fd-4428-b885-8398b6343624\") " Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.741899 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/523066ca-dde9-4160-b20f-d7e05a8f564d-config" (OuterVolumeSpecName: "config") pod "523066ca-dde9-4160-b20f-d7e05a8f564d" (UID: "523066ca-dde9-4160-b20f-d7e05a8f564d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.742377 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.742409 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523066ca-dde9-4160-b20f-d7e05a8f564d-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.742388 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "503c78af-26fd-4428-b885-8398b6343624" (UID: "503c78af-26fd-4428-b885-8398b6343624"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.781845 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/523066ca-dde9-4160-b20f-d7e05a8f564d-kube-api-access-kn9pt" (OuterVolumeSpecName: "kube-api-access-kn9pt") pod "523066ca-dde9-4160-b20f-d7e05a8f564d" (UID: "523066ca-dde9-4160-b20f-d7e05a8f564d"). InnerVolumeSpecName "kube-api-access-kn9pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.781912 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/503c78af-26fd-4428-b885-8398b6343624-kube-api-access-kqbvn" (OuterVolumeSpecName: "kube-api-access-kqbvn") pod "503c78af-26fd-4428-b885-8398b6343624" (UID: "503c78af-26fd-4428-b885-8398b6343624"). InnerVolumeSpecName "kube-api-access-kqbvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.844288 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/503c78af-26fd-4428-b885-8398b6343624-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.844331 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn9pt\" (UniqueName: \"kubernetes.io/projected/523066ca-dde9-4160-b20f-d7e05a8f564d-kube-api-access-kn9pt\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:27 crc kubenswrapper[4990]: I1203 12:58:27.844348 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqbvn\" (UniqueName: \"kubernetes.io/projected/503c78af-26fd-4428-b885-8398b6343624-kube-api-access-kqbvn\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.046334 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" event={"ID":"503c78af-26fd-4428-b885-8398b6343624","Type":"ContainerDied","Data":"f6c4105879d51148e0630fb6315a735d178966ab0c39c57dfdb3bb0087b20851"} Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.046624 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zzwqx" Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.050780 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" event={"ID":"523066ca-dde9-4160-b20f-d7e05a8f564d","Type":"ContainerDied","Data":"88aaedfd0e8f291f6711e29996a55e816a050dcff816cd774cf3c43f32339b3a"} Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.050909 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s2lj9" Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.056280 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lbt2k" event={"ID":"3f81fd15-3954-40f5-9033-18cb8e9b01ed","Type":"ContainerStarted","Data":"b7a32782e41e59915b60a00fb31c3f8d682e0e76ccae43dec041744671ccd654"} Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.062237 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d0a6648-1fc6-4928-ab9f-19829eaa0726","Type":"ContainerStarted","Data":"68b733f8c158a6602693114ab4de70f30583ca7579fb2a07014dc57a6407e951"} Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.066694 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5c414845-915f-47b6-8a52-0e6f4e9ad886","Type":"ContainerStarted","Data":"b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1"} Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.183566 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2lj9"] Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.195085 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s2lj9"] Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.213205 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zzwqx"] Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.222668 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zzwqx"] Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.281042 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="503c78af-26fd-4428-b885-8398b6343624" path="/var/lib/kubelet/pods/503c78af-26fd-4428-b885-8398b6343624/volumes" Dec 03 12:58:28 crc kubenswrapper[4990]: I1203 12:58:28.281419 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="523066ca-dde9-4160-b20f-d7e05a8f564d" path="/var/lib/kubelet/pods/523066ca-dde9-4160-b20f-d7e05a8f564d/volumes" Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.117084 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"24350652-92e4-40b6-82d5-4b785a2a0314","Type":"ContainerStarted","Data":"4d010f28e937ef21830c820152f4d668de182c70eed4c31c17ce723b919df1ce"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.118571 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.121071 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn" event={"ID":"83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5","Type":"ContainerStarted","Data":"6141b77176c738646d77d2beec780bbffb94c61f04266414846215117a29265b"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.121788 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6frjn" Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.123664 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"63ff1335-eac5-4524-b448-785488679bff","Type":"ContainerStarted","Data":"43197919466dc6cc4529d2ebb71106ef00a6aa997de588706031ed5a1f30a73e"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.125145 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8ae85bd2-714b-4a62-b287-c2a9780753bd","Type":"ContainerStarted","Data":"3fcf178d81b49f954b6cd1f564d0bf08ed0a305005d6650d6965d74bd48d8bdc"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.126669 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81e14867-9f93-4820-ba0f-c6dbd12e31db","Type":"ContainerStarted","Data":"06526eb0b6410b59349d2f2e39d60c97c5f254a5d808ba66b309762f49167ac3"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.128350 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lbt2k" event={"ID":"3f81fd15-3954-40f5-9033-18cb8e9b01ed","Type":"ContainerStarted","Data":"b78979e583de0fd31d6096068f869df1faddceb987760ace82ee9b612b171655"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.130151 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9896307b-5428-4f09-a960-1001d654e2be","Type":"ContainerStarted","Data":"b661ba261d228bce4d30aea856e638492446a4cb279683d720a5805c53459265"} Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.130670 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.143732 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=60.810323776 podStartE2EDuration="1m6.14370422s" podCreationTimestamp="2025-12-03 12:57:28 +0000 UTC" firstStartedPulling="2025-12-03 12:58:26.831704321 +0000 UTC m=+1254.973615550" lastFinishedPulling="2025-12-03 12:58:32.165084765 +0000 UTC m=+1260.306995994" observedRunningTime="2025-12-03 12:58:34.137484976 +0000 UTC m=+1262.279396225" watchObservedRunningTime="2025-12-03 12:58:34.14370422 +0000 UTC m=+1262.285615449" Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.162651 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=56.842074179 podStartE2EDuration="1m3.162628369s" podCreationTimestamp="2025-12-03 12:57:31 +0000 UTC" firstStartedPulling="2025-12-03 12:58:26.732697441 +0000 UTC m=+1254.874608670" lastFinishedPulling="2025-12-03 12:58:33.053251631 +0000 UTC m=+1261.195162860" observedRunningTime="2025-12-03 12:58:34.15620584 +0000 UTC m=+1262.298117079" watchObservedRunningTime="2025-12-03 12:58:34.162628369 +0000 UTC m=+1262.304539608" Dec 03 12:58:34 crc kubenswrapper[4990]: I1203 12:58:34.177280 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6frjn" podStartSLOduration=55.300362121 podStartE2EDuration="1m1.177265695s" podCreationTimestamp="2025-12-03 12:57:33 +0000 UTC" firstStartedPulling="2025-12-03 12:58:26.288115239 +0000 UTC m=+1254.430026478" lastFinishedPulling="2025-12-03 12:58:32.165018813 +0000 UTC m=+1260.306930052" observedRunningTime="2025-12-03 12:58:34.175547129 +0000 UTC m=+1262.317458358" watchObservedRunningTime="2025-12-03 12:58:34.177265695 +0000 UTC m=+1262.319176924" Dec 03 12:58:35 crc kubenswrapper[4990]: I1203 12:58:35.139364 4990 generic.go:334] "Generic (PLEG): container finished" podID="3f81fd15-3954-40f5-9033-18cb8e9b01ed" containerID="b78979e583de0fd31d6096068f869df1faddceb987760ace82ee9b612b171655" exitCode=0 Dec 03 12:58:35 crc kubenswrapper[4990]: I1203 12:58:35.139636 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lbt2k" event={"ID":"3f81fd15-3954-40f5-9033-18cb8e9b01ed","Type":"ContainerDied","Data":"b78979e583de0fd31d6096068f869df1faddceb987760ace82ee9b612b171655"} Dec 03 12:58:38 crc kubenswrapper[4990]: I1203 12:58:38.948008 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 12:58:39 crc kubenswrapper[4990]: I1203 12:58:39.171749 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4fb6a200-beee-4b47-ac8d-e370e307070f","Type":"ContainerStarted","Data":"d619589400fcba8f1b424ae4fbac3806186b62d9e9965a4acf56bebded55b568"} Dec 03 12:58:40 crc kubenswrapper[4990]: I1203 12:58:40.186053 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lbt2k" event={"ID":"3f81fd15-3954-40f5-9033-18cb8e9b01ed","Type":"ContainerStarted","Data":"6b73acc0dd9ecdbf08b76cdb60790aa379c6783a01bd1640e217c3652548b26d"} Dec 03 12:58:40 crc kubenswrapper[4990]: I1203 12:58:40.186498 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lbt2k" event={"ID":"3f81fd15-3954-40f5-9033-18cb8e9b01ed","Type":"ContainerStarted","Data":"fb63cb7211fc9dcbcbd395416f391486527bfa1e49372af6d5aa3a3241dde48b"} Dec 03 12:58:40 crc kubenswrapper[4990]: I1203 12:58:40.187387 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:58:40 crc kubenswrapper[4990]: I1203 12:58:40.187416 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:58:40 crc kubenswrapper[4990]: I1203 12:58:40.242326 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-lbt2k" podStartSLOduration=62.226222669 podStartE2EDuration="1m7.242305427s" podCreationTimestamp="2025-12-03 12:57:33 +0000 UTC" firstStartedPulling="2025-12-03 12:58:27.064134889 +0000 UTC m=+1255.206046118" lastFinishedPulling="2025-12-03 12:58:32.080217647 +0000 UTC m=+1260.222128876" observedRunningTime="2025-12-03 12:58:40.2385897 +0000 UTC m=+1268.380500949" watchObservedRunningTime="2025-12-03 12:58:40.242305427 +0000 UTC m=+1268.384216656" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.425576 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.512612 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-c974j"] Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.614843 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxq69"] Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.620528 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.639469 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxq69"] Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.723070 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-config\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.723124 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkrqf\" (UniqueName: \"kubernetes.io/projected/25351f52-5157-47f8-b963-dd1b56ce5a22-kube-api-access-dkrqf\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.723182 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.824225 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.824357 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-config\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.824386 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkrqf\" (UniqueName: \"kubernetes.io/projected/25351f52-5157-47f8-b963-dd1b56ce5a22-kube-api-access-dkrqf\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.825138 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.825331 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-config\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.855157 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkrqf\" (UniqueName: \"kubernetes.io/projected/25351f52-5157-47f8-b963-dd1b56ce5a22-kube-api-access-dkrqf\") pod \"dnsmasq-dns-7cb5889db5-fxq69\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:41 crc kubenswrapper[4990]: I1203 12:58:41.968576 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.706965 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.718617 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.721526 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.721653 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.721665 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.724244 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.728558 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-sm6p9" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.850425 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f7d434c5-0459-4d5a-a401-4a0c4b82a553-cache\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.850498 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.850534 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw6bt\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-kube-api-access-cw6bt\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.850595 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f7d434c5-0459-4d5a-a401-4a0c4b82a553-lock\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.850627 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.910623 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxq69"] Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952225 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f7d434c5-0459-4d5a-a401-4a0c4b82a553-cache\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952284 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952335 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw6bt\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-kube-api-access-cw6bt\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952394 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f7d434c5-0459-4d5a-a401-4a0c4b82a553-lock\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952439 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: E1203 12:58:42.952583 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:58:42 crc kubenswrapper[4990]: E1203 12:58:42.952626 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:58:42 crc kubenswrapper[4990]: E1203 12:58:42.952690 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:58:43.452664525 +0000 UTC m=+1271.594575754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952796 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952840 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f7d434c5-0459-4d5a-a401-4a0c4b82a553-cache\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.952893 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f7d434c5-0459-4d5a-a401-4a0c4b82a553-lock\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: W1203 12:58:42.964113 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25351f52_5157_47f8_b963_dd1b56ce5a22.slice/crio-aeabb04dc3cdb2026c55267d7bc272caf4109b3ccdbca9d8a77fc7ac33d53247 WatchSource:0}: Error finding container aeabb04dc3cdb2026c55267d7bc272caf4109b3ccdbca9d8a77fc7ac33d53247: Status 404 returned error can't find the container with id aeabb04dc3cdb2026c55267d7bc272caf4109b3ccdbca9d8a77fc7ac33d53247 Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.977328 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:42 crc kubenswrapper[4990]: I1203 12:58:42.998139 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw6bt\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-kube-api-access-cw6bt\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.213957 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" event={"ID":"25351f52-5157-47f8-b963-dd1b56ce5a22","Type":"ContainerStarted","Data":"aeabb04dc3cdb2026c55267d7bc272caf4109b3ccdbca9d8a77fc7ac33d53247"} Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.320870 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-gqdz2"] Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.322185 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.324625 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.325316 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.325811 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.334935 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gqdz2"] Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461032 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461120 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-etc-swift\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461156 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-dispersionconf\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461215 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-combined-ca-bundle\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461245 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-scripts\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: E1203 12:58:43.461327 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:58:43 crc kubenswrapper[4990]: E1203 12:58:43.461365 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:58:43 crc kubenswrapper[4990]: E1203 12:58:43.461415 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:58:44.461398359 +0000 UTC m=+1272.603309588 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461419 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l62kc\" (UniqueName: \"kubernetes.io/projected/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-kube-api-access-l62kc\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461563 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-ring-data-devices\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.461626 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-swiftconf\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.563274 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-ring-data-devices\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.563328 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-swiftconf\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.563416 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-etc-swift\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.564024 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-etc-swift\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.564210 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-dispersionconf\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.564267 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-combined-ca-bundle\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.564301 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-scripts\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.564337 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l62kc\" (UniqueName: \"kubernetes.io/projected/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-kube-api-access-l62kc\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.564374 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-ring-data-devices\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.565090 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-scripts\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.567460 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-swiftconf\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.567932 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-combined-ca-bundle\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.568514 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-dispersionconf\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.581418 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l62kc\" (UniqueName: \"kubernetes.io/projected/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-kube-api-access-l62kc\") pod \"swift-ring-rebalance-gqdz2\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:43 crc kubenswrapper[4990]: I1203 12:58:43.648141 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.128647 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gqdz2"] Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.226188 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8ae85bd2-714b-4a62-b287-c2a9780753bd","Type":"ContainerStarted","Data":"3b0ee91fc8898d31dd077a70682681fbe5a9d54b6a82545937186f8be6372703"} Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.228527 4990 generic.go:334] "Generic (PLEG): container finished" podID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerID="a5006df1726239cff9c557bb720ae173afdfaaf47878e383e9b290df9aab61d4" exitCode=0 Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.228617 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" event={"ID":"25351f52-5157-47f8-b963-dd1b56ce5a22","Type":"ContainerDied","Data":"a5006df1726239cff9c557bb720ae173afdfaaf47878e383e9b290df9aab61d4"} Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.230898 4990 generic.go:334] "Generic (PLEG): container finished" podID="4b643524-7647-40ed-9151-90c0921700cf" containerID="a4b4c3203f5b8e08432c6adb1b01ddf87ade610daad898f4f76276ecd17305b2" exitCode=0 Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.230981 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-c974j" event={"ID":"4b643524-7647-40ed-9151-90c0921700cf","Type":"ContainerDied","Data":"a4b4c3203f5b8e08432c6adb1b01ddf87ade610daad898f4f76276ecd17305b2"} Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.232736 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gqdz2" event={"ID":"73f3c670-ab5f-4ab9-9278-9aa008c9a35b","Type":"ContainerStarted","Data":"67c00efb09ca140e387ad553554db9a449bba770f221723a550dc65ed14162f0"} Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.239067 4990 generic.go:334] "Generic (PLEG): container finished" podID="5c617111-b3c2-431b-80eb-c3c30f0b4269" containerID="b2c4b9311b340123e0fd788b8ce50c71b4cb72a971f458970c675825de8256fa" exitCode=0 Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.239133 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" event={"ID":"5c617111-b3c2-431b-80eb-c3c30f0b4269","Type":"ContainerDied","Data":"b2c4b9311b340123e0fd788b8ce50c71b4cb72a971f458970c675825de8256fa"} Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.246540 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"63ff1335-eac5-4524-b448-785488679bff","Type":"ContainerStarted","Data":"58787f05facba0cd994ff34ba7c24b980a713c45a336c944f4d288c253ca159d"} Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.257004 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=55.207177015 podStartE2EDuration="1m11.256984924s" podCreationTimestamp="2025-12-03 12:57:33 +0000 UTC" firstStartedPulling="2025-12-03 12:58:26.926029138 +0000 UTC m=+1255.067940367" lastFinishedPulling="2025-12-03 12:58:42.975837047 +0000 UTC m=+1271.117748276" observedRunningTime="2025-12-03 12:58:44.254764645 +0000 UTC m=+1272.396675894" watchObservedRunningTime="2025-12-03 12:58:44.256984924 +0000 UTC m=+1272.398896153" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.398533 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=50.68885935 podStartE2EDuration="1m7.398509615s" podCreationTimestamp="2025-12-03 12:57:37 +0000 UTC" firstStartedPulling="2025-12-03 12:58:26.287603466 +0000 UTC m=+1254.429514715" lastFinishedPulling="2025-12-03 12:58:42.997253751 +0000 UTC m=+1271.139164980" observedRunningTime="2025-12-03 12:58:44.363112092 +0000 UTC m=+1272.505023341" watchObservedRunningTime="2025-12-03 12:58:44.398509615 +0000 UTC m=+1272.540420864" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.491863 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:44 crc kubenswrapper[4990]: E1203 12:58:44.492068 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:58:44 crc kubenswrapper[4990]: E1203 12:58:44.492277 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:58:44 crc kubenswrapper[4990]: E1203 12:58:44.492321 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:58:46.492305518 +0000 UTC m=+1274.634216737 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:58:44 crc kubenswrapper[4990]: E1203 12:58:44.517329 4990 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 03 12:58:44 crc kubenswrapper[4990]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/5c617111-b3c2-431b-80eb-c3c30f0b4269/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 12:58:44 crc kubenswrapper[4990]: > podSandboxID="6d22ce3beccef60ff39435f3a45565d6ca21ceab6952aa105d38be28befb7677" Dec 03 12:58:44 crc kubenswrapper[4990]: E1203 12:58:44.517656 4990 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 12:58:44 crc kubenswrapper[4990]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qn2zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-dh6zz_openstack(5c617111-b3c2-431b-80eb-c3c30f0b4269): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/5c617111-b3c2-431b-80eb-c3c30f0b4269/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 12:58:44 crc kubenswrapper[4990]: > logger="UnhandledError" Dec 03 12:58:44 crc kubenswrapper[4990]: E1203 12:58:44.518863 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/5c617111-b3c2-431b-80eb-c3c30f0b4269/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.540355 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.593696 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt525\" (UniqueName: \"kubernetes.io/projected/4b643524-7647-40ed-9151-90c0921700cf-kube-api-access-nt525\") pod \"4b643524-7647-40ed-9151-90c0921700cf\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.593972 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-dns-svc\") pod \"4b643524-7647-40ed-9151-90c0921700cf\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.594076 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-config\") pod \"4b643524-7647-40ed-9151-90c0921700cf\" (UID: \"4b643524-7647-40ed-9151-90c0921700cf\") " Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.599674 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b643524-7647-40ed-9151-90c0921700cf-kube-api-access-nt525" (OuterVolumeSpecName: "kube-api-access-nt525") pod "4b643524-7647-40ed-9151-90c0921700cf" (UID: "4b643524-7647-40ed-9151-90c0921700cf"). InnerVolumeSpecName "kube-api-access-nt525". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.614534 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b643524-7647-40ed-9151-90c0921700cf" (UID: "4b643524-7647-40ed-9151-90c0921700cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.617611 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-config" (OuterVolumeSpecName: "config") pod "4b643524-7647-40ed-9151-90c0921700cf" (UID: "4b643524-7647-40ed-9151-90c0921700cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.695752 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt525\" (UniqueName: \"kubernetes.io/projected/4b643524-7647-40ed-9151-90c0921700cf-kube-api-access-nt525\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.695787 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.695796 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b643524-7647-40ed-9151-90c0921700cf-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.881879 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 12:58:44 crc kubenswrapper[4990]: I1203 12:58:44.932690 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.076548 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.257620 4990 generic.go:334] "Generic (PLEG): container finished" podID="81e14867-9f93-4820-ba0f-c6dbd12e31db" containerID="06526eb0b6410b59349d2f2e39d60c97c5f254a5d808ba66b309762f49167ac3" exitCode=0 Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.257692 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81e14867-9f93-4820-ba0f-c6dbd12e31db","Type":"ContainerDied","Data":"06526eb0b6410b59349d2f2e39d60c97c5f254a5d808ba66b309762f49167ac3"} Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.266283 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-c974j" event={"ID":"4b643524-7647-40ed-9151-90c0921700cf","Type":"ContainerDied","Data":"fa2cc7f824400945b997133b70aa02f0c00943ff7d7322b4fe10303844885144"} Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.266309 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-c974j" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.266336 4990 scope.go:117] "RemoveContainer" containerID="a4b4c3203f5b8e08432c6adb1b01ddf87ade610daad898f4f76276ecd17305b2" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.269395 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" event={"ID":"25351f52-5157-47f8-b963-dd1b56ce5a22","Type":"ContainerStarted","Data":"4fd7235ef2951fa44c32f912cca0258424f0d4ec8bf7af115ed1fa97eecc03ba"} Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.270110 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.314897 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" podStartSLOduration=4.314877554 podStartE2EDuration="4.314877554s" podCreationTimestamp="2025-12-03 12:58:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:58:45.310438197 +0000 UTC m=+1273.452349426" watchObservedRunningTime="2025-12-03 12:58:45.314877554 +0000 UTC m=+1273.456788783" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.344507 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.466563 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-c974j"] Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.478056 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-c974j"] Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.576871 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dh6zz"] Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.596539 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-fwm6p"] Dec 03 12:58:45 crc kubenswrapper[4990]: E1203 12:58:45.597222 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b643524-7647-40ed-9151-90c0921700cf" containerName="init" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.597249 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b643524-7647-40ed-9151-90c0921700cf" containerName="init" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.597479 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b643524-7647-40ed-9151-90c0921700cf" containerName="init" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.600862 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.604678 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.613328 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-fwm6p"] Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.617741 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b98cg\" (UniqueName: \"kubernetes.io/projected/4167e405-5683-46fe-a8bb-f098143088ac-kube-api-access-b98cg\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.617815 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-ovsdbserver-sb\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.617915 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-dns-svc\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.617947 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-config\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.697024 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-dpmrl"] Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.698473 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.704964 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.716809 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-dpmrl"] Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.720999 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb4wz\" (UniqueName: \"kubernetes.io/projected/caff07ff-f45e-4438-96a7-545d4de585c0-kube-api-access-mb4wz\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721088 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caff07ff-f45e-4438-96a7-545d4de585c0-config\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721117 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/caff07ff-f45e-4438-96a7-545d4de585c0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721163 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/caff07ff-f45e-4438-96a7-545d4de585c0-ovn-rundir\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721196 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-dns-svc\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721232 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-config\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721348 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b98cg\" (UniqueName: \"kubernetes.io/projected/4167e405-5683-46fe-a8bb-f098143088ac-kube-api-access-b98cg\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721374 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/caff07ff-f45e-4438-96a7-545d4de585c0-ovs-rundir\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721404 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caff07ff-f45e-4438-96a7-545d4de585c0-combined-ca-bundle\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.721472 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-ovsdbserver-sb\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.722763 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-dns-svc\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.723721 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-config\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.731605 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-ovsdbserver-sb\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.744892 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b98cg\" (UniqueName: \"kubernetes.io/projected/4167e405-5683-46fe-a8bb-f098143088ac-kube-api-access-b98cg\") pod \"dnsmasq-dns-8cc7fc4dc-fwm6p\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.822752 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/caff07ff-f45e-4438-96a7-545d4de585c0-ovs-rundir\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.822810 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caff07ff-f45e-4438-96a7-545d4de585c0-combined-ca-bundle\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.822876 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb4wz\" (UniqueName: \"kubernetes.io/projected/caff07ff-f45e-4438-96a7-545d4de585c0-kube-api-access-mb4wz\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.822916 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caff07ff-f45e-4438-96a7-545d4de585c0-config\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.822936 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/caff07ff-f45e-4438-96a7-545d4de585c0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.822972 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/caff07ff-f45e-4438-96a7-545d4de585c0-ovn-rundir\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.823104 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/caff07ff-f45e-4438-96a7-545d4de585c0-ovs-rundir\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.823124 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/caff07ff-f45e-4438-96a7-545d4de585c0-ovn-rundir\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.823886 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caff07ff-f45e-4438-96a7-545d4de585c0-config\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.829421 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/caff07ff-f45e-4438-96a7-545d4de585c0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.831596 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caff07ff-f45e-4438-96a7-545d4de585c0-combined-ca-bundle\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.843798 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb4wz\" (UniqueName: \"kubernetes.io/projected/caff07ff-f45e-4438-96a7-545d4de585c0-kube-api-access-mb4wz\") pod \"ovn-controller-metrics-dpmrl\" (UID: \"caff07ff-f45e-4438-96a7-545d4de585c0\") " pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:45 crc kubenswrapper[4990]: I1203 12:58:45.933429 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.027763 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-dpmrl" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.098955 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxq69"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.168277 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bhcb6"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.176349 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.188386 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bhcb6"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.196258 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.200375 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.245225 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-config\") pod \"5c617111-b3c2-431b-80eb-c3c30f0b4269\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.245358 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-dns-svc\") pod \"5c617111-b3c2-431b-80eb-c3c30f0b4269\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.245432 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn2zc\" (UniqueName: \"kubernetes.io/projected/5c617111-b3c2-431b-80eb-c3c30f0b4269-kube-api-access-qn2zc\") pod \"5c617111-b3c2-431b-80eb-c3c30f0b4269\" (UID: \"5c617111-b3c2-431b-80eb-c3c30f0b4269\") " Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.246521 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.246685 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfcr7\" (UniqueName: \"kubernetes.io/projected/3186fb70-d995-467f-beef-04cd48564ecc-kube-api-access-gfcr7\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.246813 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-config\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.246875 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.246932 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.265413 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c617111-b3c2-431b-80eb-c3c30f0b4269-kube-api-access-qn2zc" (OuterVolumeSpecName: "kube-api-access-qn2zc") pod "5c617111-b3c2-431b-80eb-c3c30f0b4269" (UID: "5c617111-b3c2-431b-80eb-c3c30f0b4269"). InnerVolumeSpecName "kube-api-access-qn2zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.313078 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b643524-7647-40ed-9151-90c0921700cf" path="/var/lib/kubelet/pods/4b643524-7647-40ed-9151-90c0921700cf/volumes" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.326299 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" event={"ID":"5c617111-b3c2-431b-80eb-c3c30f0b4269","Type":"ContainerDied","Data":"6d22ce3beccef60ff39435f3a45565d6ca21ceab6952aa105d38be28befb7677"} Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.326375 4990 scope.go:117] "RemoveContainer" containerID="b2c4b9311b340123e0fd788b8ce50c71b4cb72a971f458970c675825de8256fa" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.326314 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dh6zz" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.336908 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.339360 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-config" (OuterVolumeSpecName: "config") pod "5c617111-b3c2-431b-80eb-c3c30f0b4269" (UID: "5c617111-b3c2-431b-80eb-c3c30f0b4269"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.348891 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.348950 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.349035 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfcr7\" (UniqueName: \"kubernetes.io/projected/3186fb70-d995-467f-beef-04cd48564ecc-kube-api-access-gfcr7\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.349121 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-config\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.349158 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.349234 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.349245 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn2zc\" (UniqueName: \"kubernetes.io/projected/5c617111-b3c2-431b-80eb-c3c30f0b4269-kube-api-access-qn2zc\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.350491 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.350997 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-config\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.351215 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.351748 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.359779 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c617111-b3c2-431b-80eb-c3c30f0b4269" (UID: "5c617111-b3c2-431b-80eb-c3c30f0b4269"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.375281 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfcr7\" (UniqueName: \"kubernetes.io/projected/3186fb70-d995-467f-beef-04cd48564ecc-kube-api-access-gfcr7\") pod \"dnsmasq-dns-b8fbc5445-bhcb6\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.451200 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c617111-b3c2-431b-80eb-c3c30f0b4269-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.544198 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.552294 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:46 crc kubenswrapper[4990]: E1203 12:58:46.553928 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:58:46 crc kubenswrapper[4990]: E1203 12:58:46.553946 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:58:46 crc kubenswrapper[4990]: E1203 12:58:46.553986 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:58:50.553973033 +0000 UTC m=+1278.695884262 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.601925 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-fwm6p"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.699498 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-dpmrl"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.731011 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dh6zz"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.740204 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dh6zz"] Dec 03 12:58:46 crc kubenswrapper[4990]: I1203 12:58:46.829584 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bhcb6"] Dec 03 12:58:46 crc kubenswrapper[4990]: W1203 12:58:46.839381 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3186fb70_d995_467f_beef_04cd48564ecc.slice/crio-a29a0650adbf1ef931a22cbf8dc4d894d0faa7cb9e7b520e581d0f96d9903429 WatchSource:0}: Error finding container a29a0650adbf1ef931a22cbf8dc4d894d0faa7cb9e7b520e581d0f96d9903429: Status 404 returned error can't find the container with id a29a0650adbf1ef931a22cbf8dc4d894d0faa7cb9e7b520e581d0f96d9903429 Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.076357 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.122191 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.343118 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" event={"ID":"4167e405-5683-46fe-a8bb-f098143088ac","Type":"ContainerStarted","Data":"96ba3517606911ed9d4ffb06c65882ff3336b7c843d6bedb24f1f5d934f072b6"} Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.346642 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81e14867-9f93-4820-ba0f-c6dbd12e31db","Type":"ContainerStarted","Data":"143fc6184fdb18dfe10b8d4c28d39c4d4ae057ad06182a9e21142a6368ba2f38"} Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.348193 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-dpmrl" event={"ID":"caff07ff-f45e-4438-96a7-545d4de585c0","Type":"ContainerStarted","Data":"62f8c6bda023d1ed40e9255f71d96b96e04d80747e65fd43e1ef9e0af92cb99d"} Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.349989 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" event={"ID":"3186fb70-d995-467f-beef-04cd48564ecc","Type":"ContainerStarted","Data":"a29a0650adbf1ef931a22cbf8dc4d894d0faa7cb9e7b520e581d0f96d9903429"} Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.350288 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="dnsmasq-dns" containerID="cri-o://4fd7235ef2951fa44c32f912cca0258424f0d4ec8bf7af115ed1fa97eecc03ba" gracePeriod=10 Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.375789 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=75.015876407 podStartE2EDuration="1m20.375771679s" podCreationTimestamp="2025-12-03 12:57:27 +0000 UTC" firstStartedPulling="2025-12-03 12:58:26.92079482 +0000 UTC m=+1255.062706049" lastFinishedPulling="2025-12-03 12:58:32.280690092 +0000 UTC m=+1260.422601321" observedRunningTime="2025-12-03 12:58:47.366056443 +0000 UTC m=+1275.507967672" watchObservedRunningTime="2025-12-03 12:58:47.375771679 +0000 UTC m=+1275.517682908" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.396051 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.705639 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 12:58:47 crc kubenswrapper[4990]: E1203 12:58:47.706250 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" containerName="init" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.706274 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" containerName="init" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.706717 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" containerName="init" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.710180 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.714894 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.715508 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-q5fnb" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.715719 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.720592 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.735900 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.885372 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-scripts\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.885427 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.885789 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pmfw\" (UniqueName: \"kubernetes.io/projected/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-kube-api-access-8pmfw\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.885841 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.885872 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-config\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.886025 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.886155 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988138 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988212 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-scripts\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988237 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988287 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pmfw\" (UniqueName: \"kubernetes.io/projected/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-kube-api-access-8pmfw\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988326 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988353 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-config\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.988377 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.990209 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-scripts\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.990356 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-config\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.990827 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.993972 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:47 crc kubenswrapper[4990]: I1203 12:58:47.994095 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.001932 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.005168 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pmfw\" (UniqueName: \"kubernetes.io/projected/7d581ac6-ba40-4773-9d18-5aa2fcdfad62-kube-api-access-8pmfw\") pod \"ovn-northd-0\" (UID: \"7d581ac6-ba40-4773-9d18-5aa2fcdfad62\") " pod="openstack/ovn-northd-0" Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.032122 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.286479 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c617111-b3c2-431b-80eb-c3c30f0b4269" path="/var/lib/kubelet/pods/5c617111-b3c2-431b-80eb-c3c30f0b4269/volumes" Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.377535 4990 generic.go:334] "Generic (PLEG): container finished" podID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerID="4fd7235ef2951fa44c32f912cca0258424f0d4ec8bf7af115ed1fa97eecc03ba" exitCode=0 Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.377618 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" event={"ID":"25351f52-5157-47f8-b963-dd1b56ce5a22","Type":"ContainerDied","Data":"4fd7235ef2951fa44c32f912cca0258424f0d4ec8bf7af115ed1fa97eecc03ba"} Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.380301 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-dpmrl" event={"ID":"caff07ff-f45e-4438-96a7-545d4de585c0","Type":"ContainerStarted","Data":"08cb33363af0cd50caf246b8633c49e6d1201bdd30af08fdaa40c2d56d4bdb3e"} Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.393819 4990 generic.go:334] "Generic (PLEG): container finished" podID="3186fb70-d995-467f-beef-04cd48564ecc" containerID="669c2fd5365603e7961c038baecd37544f799c4ee0a2fe0a7b85a05baf77510f" exitCode=0 Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.393949 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" event={"ID":"3186fb70-d995-467f-beef-04cd48564ecc","Type":"ContainerDied","Data":"669c2fd5365603e7961c038baecd37544f799c4ee0a2fe0a7b85a05baf77510f"} Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.396382 4990 generic.go:334] "Generic (PLEG): container finished" podID="4167e405-5683-46fe-a8bb-f098143088ac" containerID="8c903b2a2c9c457c79f1b01f606c38db2ff4273852ccd5c81a063266e32cb71d" exitCode=0 Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.397357 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" event={"ID":"4167e405-5683-46fe-a8bb-f098143088ac","Type":"ContainerDied","Data":"8c903b2a2c9c457c79f1b01f606c38db2ff4273852ccd5c81a063266e32cb71d"} Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.409374 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-dpmrl" podStartSLOduration=3.409350779 podStartE2EDuration="3.409350779s" podCreationTimestamp="2025-12-03 12:58:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:58:48.402616582 +0000 UTC m=+1276.544527821" watchObservedRunningTime="2025-12-03 12:58:48.409350779 +0000 UTC m=+1276.551262008" Dec 03 12:58:48 crc kubenswrapper[4990]: I1203 12:58:48.564435 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 12:58:49 crc kubenswrapper[4990]: I1203 12:58:49.099177 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 12:58:49 crc kubenswrapper[4990]: I1203 12:58:49.099257 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 12:58:50 crc kubenswrapper[4990]: I1203 12:58:50.651649 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:50 crc kubenswrapper[4990]: E1203 12:58:50.652038 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:58:50 crc kubenswrapper[4990]: E1203 12:58:50.652279 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:58:50 crc kubenswrapper[4990]: E1203 12:58:50.652389 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:58:58.652356045 +0000 UTC m=+1286.794267314 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:58:53 crc kubenswrapper[4990]: I1203 12:58:53.439023 4990 generic.go:334] "Generic (PLEG): container finished" podID="4fb6a200-beee-4b47-ac8d-e370e307070f" containerID="d619589400fcba8f1b424ae4fbac3806186b62d9e9965a4acf56bebded55b568" exitCode=0 Dec 03 12:58:53 crc kubenswrapper[4990]: I1203 12:58:53.439107 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4fb6a200-beee-4b47-ac8d-e370e307070f","Type":"ContainerDied","Data":"d619589400fcba8f1b424ae4fbac3806186b62d9e9965a4acf56bebded55b568"} Dec 03 12:58:56 crc kubenswrapper[4990]: I1203 12:58:56.969909 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.107:5353: i/o timeout" Dec 03 12:58:58 crc kubenswrapper[4990]: I1203 12:58:58.697699 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:58:58 crc kubenswrapper[4990]: E1203 12:58:58.697997 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:58:58 crc kubenswrapper[4990]: E1203 12:58:58.698472 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:58:58 crc kubenswrapper[4990]: E1203 12:58:58.698543 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:59:14.698524662 +0000 UTC m=+1302.840435891 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:58:59 crc kubenswrapper[4990]: W1203 12:58:59.363104 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d581ac6_ba40_4773_9d18_5aa2fcdfad62.slice/crio-1d20ce2e008584bca855b77b932eabe2bd2d3323f18c31146ceddd298a496245 WatchSource:0}: Error finding container 1d20ce2e008584bca855b77b932eabe2bd2d3323f18c31146ceddd298a496245: Status 404 returned error can't find the container with id 1d20ce2e008584bca855b77b932eabe2bd2d3323f18c31146ceddd298a496245 Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.488101 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7d581ac6-ba40-4773-9d18-5aa2fcdfad62","Type":"ContainerStarted","Data":"1d20ce2e008584bca855b77b932eabe2bd2d3323f18c31146ceddd298a496245"} Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.491189 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" event={"ID":"25351f52-5157-47f8-b963-dd1b56ce5a22","Type":"ContainerDied","Data":"aeabb04dc3cdb2026c55267d7bc272caf4109b3ccdbca9d8a77fc7ac33d53247"} Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.491333 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeabb04dc3cdb2026c55267d7bc272caf4109b3ccdbca9d8a77fc7ac33d53247" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.519511 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.614281 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-config\") pod \"25351f52-5157-47f8-b963-dd1b56ce5a22\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.614436 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkrqf\" (UniqueName: \"kubernetes.io/projected/25351f52-5157-47f8-b963-dd1b56ce5a22-kube-api-access-dkrqf\") pod \"25351f52-5157-47f8-b963-dd1b56ce5a22\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.614586 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-dns-svc\") pod \"25351f52-5157-47f8-b963-dd1b56ce5a22\" (UID: \"25351f52-5157-47f8-b963-dd1b56ce5a22\") " Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.620401 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25351f52-5157-47f8-b963-dd1b56ce5a22-kube-api-access-dkrqf" (OuterVolumeSpecName: "kube-api-access-dkrqf") pod "25351f52-5157-47f8-b963-dd1b56ce5a22" (UID: "25351f52-5157-47f8-b963-dd1b56ce5a22"). InnerVolumeSpecName "kube-api-access-dkrqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.668038 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "25351f52-5157-47f8-b963-dd1b56ce5a22" (UID: "25351f52-5157-47f8-b963-dd1b56ce5a22"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.689861 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-config" (OuterVolumeSpecName: "config") pod "25351f52-5157-47f8-b963-dd1b56ce5a22" (UID: "25351f52-5157-47f8-b963-dd1b56ce5a22"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.716400 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.717571 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkrqf\" (UniqueName: \"kubernetes.io/projected/25351f52-5157-47f8-b963-dd1b56ce5a22-kube-api-access-dkrqf\") on node \"crc\" DevicePath \"\"" Dec 03 12:58:59 crc kubenswrapper[4990]: I1203 12:58:59.718087 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25351f52-5157-47f8-b963-dd1b56ce5a22-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:00 crc kubenswrapper[4990]: I1203 12:59:00.500427 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" Dec 03 12:59:00 crc kubenswrapper[4990]: I1203 12:59:00.534320 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxq69"] Dec 03 12:59:00 crc kubenswrapper[4990]: I1203 12:59:00.546812 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxq69"] Dec 03 12:59:01 crc kubenswrapper[4990]: I1203 12:59:01.512930 4990 generic.go:334] "Generic (PLEG): container finished" podID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerID="68b733f8c158a6602693114ab4de70f30583ca7579fb2a07014dc57a6407e951" exitCode=0 Dec 03 12:59:01 crc kubenswrapper[4990]: I1203 12:59:01.513005 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d0a6648-1fc6-4928-ab9f-19829eaa0726","Type":"ContainerDied","Data":"68b733f8c158a6602693114ab4de70f30583ca7579fb2a07014dc57a6407e951"} Dec 03 12:59:01 crc kubenswrapper[4990]: I1203 12:59:01.514760 4990 generic.go:334] "Generic (PLEG): container finished" podID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerID="b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1" exitCode=0 Dec 03 12:59:01 crc kubenswrapper[4990]: I1203 12:59:01.514799 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5c414845-915f-47b6-8a52-0e6f4e9ad886","Type":"ContainerDied","Data":"b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1"} Dec 03 12:59:01 crc kubenswrapper[4990]: I1203 12:59:01.971439 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-fxq69" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.107:5353: i/o timeout" Dec 03 12:59:02 crc kubenswrapper[4990]: I1203 12:59:02.276641 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" path="/var/lib/kubelet/pods/25351f52-5157-47f8-b963-dd1b56ce5a22/volumes" Dec 03 12:59:03 crc kubenswrapper[4990]: I1203 12:59:03.286512 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:59:03 crc kubenswrapper[4990]: I1203 12:59:03.286577 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:59:03 crc kubenswrapper[4990]: I1203 12:59:03.536408 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" event={"ID":"3186fb70-d995-467f-beef-04cd48564ecc","Type":"ContainerStarted","Data":"9cf5a63ce8640f3d8e377ccce33ba35447c2751cdaa09f8217bf45f861b64f85"} Dec 03 12:59:03 crc kubenswrapper[4990]: I1203 12:59:03.538970 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" event={"ID":"4167e405-5683-46fe-a8bb-f098143088ac","Type":"ContainerStarted","Data":"d890669d3f2aec491fdaf9f0db83cd94c93ccc53c269f62091b0f31d9aad546d"} Dec 03 12:59:04 crc kubenswrapper[4990]: I1203 12:59:04.198670 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6frjn" podUID="83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5" containerName="ovn-controller" probeResult="failure" output=< Dec 03 12:59:04 crc kubenswrapper[4990]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 12:59:04 crc kubenswrapper[4990]: > Dec 03 12:59:04 crc kubenswrapper[4990]: I1203 12:59:04.549126 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4fb6a200-beee-4b47-ac8d-e370e307070f","Type":"ContainerStarted","Data":"f601e06b2d944fce2d0dbca7d5595949e44bab0c0c553c8dff756403bee95066"} Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.202680 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6frjn" podUID="83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5" containerName="ovn-controller" probeResult="failure" output=< Dec 03 12:59:09 crc kubenswrapper[4990]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 12:59:09 crc kubenswrapper[4990]: > Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.246277 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.264347 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lbt2k" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.506803 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6frjn-config-cdn9n"] Dec 03 12:59:09 crc kubenswrapper[4990]: E1203 12:59:09.507215 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="init" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.507239 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="init" Dec 03 12:59:09 crc kubenswrapper[4990]: E1203 12:59:09.507278 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="dnsmasq-dns" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.507289 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="dnsmasq-dns" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.507519 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="25351f52-5157-47f8-b963-dd1b56ce5a22" containerName="dnsmasq-dns" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.508210 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.510806 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.517751 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6frjn-config-cdn9n"] Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.592130 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d0a6648-1fc6-4928-ab9f-19829eaa0726","Type":"ContainerStarted","Data":"b6657e9f631e4d45ce3e912956fa8ee4bfd97492332688531321b57488cbbc45"} Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.594728 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5c414845-915f-47b6-8a52-0e6f4e9ad886","Type":"ContainerStarted","Data":"b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413"} Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.595085 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.598587 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.611599 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-log-ovn\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.611643 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-additional-scripts\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.611705 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run-ovn\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.611745 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-scripts\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.611809 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.611837 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7vzs\" (UniqueName: \"kubernetes.io/projected/8979791b-6642-4783-95a9-02a58fcaa400-kube-api-access-b7vzs\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.625157 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" podStartSLOduration=24.625141701 podStartE2EDuration="24.625141701s" podCreationTimestamp="2025-12-03 12:58:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:09.620835327 +0000 UTC m=+1297.762746556" watchObservedRunningTime="2025-12-03 12:59:09.625141701 +0000 UTC m=+1297.767052930" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714008 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-log-ovn\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714070 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-additional-scripts\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714143 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run-ovn\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714200 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-scripts\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714275 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714311 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7vzs\" (UniqueName: \"kubernetes.io/projected/8979791b-6642-4783-95a9-02a58fcaa400-kube-api-access-b7vzs\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714792 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.714861 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run-ovn\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.715211 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-additional-scripts\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.716289 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-scripts\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.716357 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-log-ovn\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.737774 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7vzs\" (UniqueName: \"kubernetes.io/projected/8979791b-6642-4783-95a9-02a58fcaa400-kube-api-access-b7vzs\") pod \"ovn-controller-6frjn-config-cdn9n\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:09 crc kubenswrapper[4990]: I1203 12:59:09.825347 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:10 crc kubenswrapper[4990]: E1203 12:59:10.064139 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified" Dec 03 12:59:10 crc kubenswrapper[4990]: E1203 12:59:10.064602 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:swift-ring-rebalance,Image:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,Command:[/usr/local/bin/swift-ring-tool all],Args:[],WorkingDir:/etc/swift,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CM_NAME,Value:swift-ring-files,ValueFrom:nil,},EnvVar{Name:NAMESPACE,Value:openstack,ValueFrom:nil,},EnvVar{Name:OWNER_APIVERSION,Value:swift.openstack.org/v1beta1,ValueFrom:nil,},EnvVar{Name:OWNER_KIND,Value:SwiftRing,ValueFrom:nil,},EnvVar{Name:OWNER_NAME,Value:swift-ring,ValueFrom:nil,},EnvVar{Name:OWNER_UID,Value:1c68715d-79f7-47ce-ae2f-82c3119dca87,ValueFrom:nil,},EnvVar{Name:SWIFT_MIN_PART_HOURS,Value:1,ValueFrom:nil,},EnvVar{Name:SWIFT_PART_POWER,Value:10,ValueFrom:nil,},EnvVar{Name:SWIFT_REPLICAS,Value:1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/swift-ring-tool,SubPath:swift-ring-tool,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:swiftconf,ReadOnly:true,MountPath:/etc/swift/swift.conf,SubPath:swift.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ring-data-devices,ReadOnly:true,MountPath:/var/lib/config-data/ring-devices,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dispersionconf,ReadOnly:true,MountPath:/etc/swift/dispersion.conf,SubPath:dispersion.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l62kc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-ring-rebalance-gqdz2_openstack(73f3c670-ab5f-4ab9-9278-9aa008c9a35b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 12:59:10 crc kubenswrapper[4990]: E1203 12:59:10.065840 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/swift-ring-rebalance-gqdz2" podUID="73f3c670-ab5f-4ab9-9278-9aa008c9a35b" Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.602334 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.605668 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 12:59:10 crc kubenswrapper[4990]: E1203 12:59:10.606680 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified\\\"\"" pod="openstack/swift-ring-rebalance-gqdz2" podUID="73f3c670-ab5f-4ab9-9278-9aa008c9a35b" Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.669650 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podStartSLOduration=24.669628668 podStartE2EDuration="24.669628668s" podCreationTimestamp="2025-12-03 12:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:10.641084646 +0000 UTC m=+1298.782995875" watchObservedRunningTime="2025-12-03 12:59:10.669628668 +0000 UTC m=+1298.811539897" Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.673219 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371931.181576 podStartE2EDuration="1m45.673199353s" podCreationTimestamp="2025-12-03 12:57:25 +0000 UTC" firstStartedPulling="2025-12-03 12:57:27.857781818 +0000 UTC m=+1195.999693037" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:10.671403725 +0000 UTC m=+1298.813314964" watchObservedRunningTime="2025-12-03 12:59:10.673199353 +0000 UTC m=+1298.815110582" Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.754870 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-fwm6p"] Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.774169 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=47.71241717 podStartE2EDuration="1m46.774049822s" podCreationTimestamp="2025-12-03 12:57:24 +0000 UTC" firstStartedPulling="2025-12-03 12:57:26.290403148 +0000 UTC m=+1194.432314377" lastFinishedPulling="2025-12-03 12:58:25.3520358 +0000 UTC m=+1253.493947029" observedRunningTime="2025-12-03 12:59:10.768315491 +0000 UTC m=+1298.910226720" watchObservedRunningTime="2025-12-03 12:59:10.774049822 +0000 UTC m=+1298.915961051" Dec 03 12:59:10 crc kubenswrapper[4990]: I1203 12:59:10.811231 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=46.499075632 podStartE2EDuration="1m46.811204731s" podCreationTimestamp="2025-12-03 12:57:24 +0000 UTC" firstStartedPulling="2025-12-03 12:57:25.972775166 +0000 UTC m=+1194.114686395" lastFinishedPulling="2025-12-03 12:58:26.284904255 +0000 UTC m=+1254.426815494" observedRunningTime="2025-12-03 12:59:10.804401851 +0000 UTC m=+1298.946313090" watchObservedRunningTime="2025-12-03 12:59:10.811204731 +0000 UTC m=+1298.953115960" Dec 03 12:59:11 crc kubenswrapper[4990]: I1203 12:59:11.477345 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6frjn-config-cdn9n"] Dec 03 12:59:11 crc kubenswrapper[4990]: I1203 12:59:11.633400 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-cdn9n" event={"ID":"8979791b-6642-4783-95a9-02a58fcaa400","Type":"ContainerStarted","Data":"3bea744b232fdced247ddae349455b84836276df8029c94bd2f60d375dfd0731"} Dec 03 12:59:11 crc kubenswrapper[4990]: I1203 12:59:11.638097 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7d581ac6-ba40-4773-9d18-5aa2fcdfad62","Type":"ContainerStarted","Data":"c5a97858d0b770c947e0335d18adbeff9884bad3c5f80c14f85c6ed8ddcf03c3"} Dec 03 12:59:12 crc kubenswrapper[4990]: I1203 12:59:12.649517 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7d581ac6-ba40-4773-9d18-5aa2fcdfad62","Type":"ContainerStarted","Data":"b7c7b755941a03c9f52a4a0640a01fb900139c3a1eccf41278dd0fbd8699015f"} Dec 03 12:59:12 crc kubenswrapper[4990]: I1203 12:59:12.649877 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 12:59:12 crc kubenswrapper[4990]: I1203 12:59:12.652495 4990 generic.go:334] "Generic (PLEG): container finished" podID="8979791b-6642-4783-95a9-02a58fcaa400" containerID="76daed6ff0ab3a3936010c09a4a3f3f32db810f207331e602d475b18359cb909" exitCode=0 Dec 03 12:59:12 crc kubenswrapper[4990]: I1203 12:59:12.652701 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" podUID="4167e405-5683-46fe-a8bb-f098143088ac" containerName="dnsmasq-dns" containerID="cri-o://d890669d3f2aec491fdaf9f0db83cd94c93ccc53c269f62091b0f31d9aad546d" gracePeriod=10 Dec 03 12:59:12 crc kubenswrapper[4990]: I1203 12:59:12.652837 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-cdn9n" event={"ID":"8979791b-6642-4783-95a9-02a58fcaa400","Type":"ContainerDied","Data":"76daed6ff0ab3a3936010c09a4a3f3f32db810f207331e602d475b18359cb909"} Dec 03 12:59:12 crc kubenswrapper[4990]: I1203 12:59:12.675696 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=14.118526285 podStartE2EDuration="25.675677477s" podCreationTimestamp="2025-12-03 12:58:47 +0000 UTC" firstStartedPulling="2025-12-03 12:58:59.367371176 +0000 UTC m=+1287.509282405" lastFinishedPulling="2025-12-03 12:59:10.924522358 +0000 UTC m=+1299.066433597" observedRunningTime="2025-12-03 12:59:12.672029071 +0000 UTC m=+1300.813940300" watchObservedRunningTime="2025-12-03 12:59:12.675677477 +0000 UTC m=+1300.817588706" Dec 03 12:59:13 crc kubenswrapper[4990]: I1203 12:59:13.662227 4990 generic.go:334] "Generic (PLEG): container finished" podID="4167e405-5683-46fe-a8bb-f098143088ac" containerID="d890669d3f2aec491fdaf9f0db83cd94c93ccc53c269f62091b0f31d9aad546d" exitCode=0 Dec 03 12:59:13 crc kubenswrapper[4990]: I1203 12:59:13.663571 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" event={"ID":"4167e405-5683-46fe-a8bb-f098143088ac","Type":"ContainerDied","Data":"d890669d3f2aec491fdaf9f0db83cd94c93ccc53c269f62091b0f31d9aad546d"} Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.011164 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106133 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run-ovn\") pod \"8979791b-6642-4783-95a9-02a58fcaa400\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106219 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-scripts\") pod \"8979791b-6642-4783-95a9-02a58fcaa400\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106263 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8979791b-6642-4783-95a9-02a58fcaa400" (UID: "8979791b-6642-4783-95a9-02a58fcaa400"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106305 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7vzs\" (UniqueName: \"kubernetes.io/projected/8979791b-6642-4783-95a9-02a58fcaa400-kube-api-access-b7vzs\") pod \"8979791b-6642-4783-95a9-02a58fcaa400\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106494 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-additional-scripts\") pod \"8979791b-6642-4783-95a9-02a58fcaa400\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106591 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-log-ovn\") pod \"8979791b-6642-4783-95a9-02a58fcaa400\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.106651 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run\") pod \"8979791b-6642-4783-95a9-02a58fcaa400\" (UID: \"8979791b-6642-4783-95a9-02a58fcaa400\") " Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.107174 4990 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.107243 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run" (OuterVolumeSpecName: "var-run") pod "8979791b-6642-4783-95a9-02a58fcaa400" (UID: "8979791b-6642-4783-95a9-02a58fcaa400"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.107287 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8979791b-6642-4783-95a9-02a58fcaa400" (UID: "8979791b-6642-4783-95a9-02a58fcaa400"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.107474 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8979791b-6642-4783-95a9-02a58fcaa400" (UID: "8979791b-6642-4783-95a9-02a58fcaa400"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.108496 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-scripts" (OuterVolumeSpecName: "scripts") pod "8979791b-6642-4783-95a9-02a58fcaa400" (UID: "8979791b-6642-4783-95a9-02a58fcaa400"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.114238 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8979791b-6642-4783-95a9-02a58fcaa400-kube-api-access-b7vzs" (OuterVolumeSpecName: "kube-api-access-b7vzs") pod "8979791b-6642-4783-95a9-02a58fcaa400" (UID: "8979791b-6642-4783-95a9-02a58fcaa400"). InnerVolumeSpecName "kube-api-access-b7vzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.179323 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6frjn" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.211230 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7vzs\" (UniqueName: \"kubernetes.io/projected/8979791b-6642-4783-95a9-02a58fcaa400-kube-api-access-b7vzs\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.211275 4990 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.211285 4990 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.211295 4990 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8979791b-6642-4783-95a9-02a58fcaa400-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.211303 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8979791b-6642-4783-95a9-02a58fcaa400-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.670050 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-cdn9n" event={"ID":"8979791b-6642-4783-95a9-02a58fcaa400","Type":"ContainerDied","Data":"3bea744b232fdced247ddae349455b84836276df8029c94bd2f60d375dfd0731"} Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.670093 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bea744b232fdced247ddae349455b84836276df8029c94bd2f60d375dfd0731" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.670200 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-cdn9n" Dec 03 12:59:14 crc kubenswrapper[4990]: I1203 12:59:14.719880 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:59:14 crc kubenswrapper[4990]: E1203 12:59:14.720020 4990 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 12:59:14 crc kubenswrapper[4990]: E1203 12:59:14.720039 4990 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 12:59:14 crc kubenswrapper[4990]: E1203 12:59:14.720088 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift podName:f7d434c5-0459-4d5a-a401-4a0c4b82a553 nodeName:}" failed. No retries permitted until 2025-12-03 12:59:46.720073457 +0000 UTC m=+1334.861984686 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift") pod "swift-storage-0" (UID: "f7d434c5-0459-4d5a-a401-4a0c4b82a553") : configmap "swift-ring-files" not found Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.133687 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6frjn-config-cdn9n"] Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.141952 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6frjn-config-cdn9n"] Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.175271 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6frjn-config-prsrj"] Dec 03 12:59:15 crc kubenswrapper[4990]: E1203 12:59:15.175752 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8979791b-6642-4783-95a9-02a58fcaa400" containerName="ovn-config" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.175780 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8979791b-6642-4783-95a9-02a58fcaa400" containerName="ovn-config" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.176011 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8979791b-6642-4783-95a9-02a58fcaa400" containerName="ovn-config" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.177021 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.184272 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.185185 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6frjn-config-prsrj"] Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.329872 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-log-ovn\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.330187 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.330223 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-scripts\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.330241 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run-ovn\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.330264 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49r5j\" (UniqueName: \"kubernetes.io/projected/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-kube-api-access-49r5j\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.330759 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-additional-scripts\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.385768 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432532 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-additional-scripts\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432601 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-log-ovn\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432655 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432702 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-scripts\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432720 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run-ovn\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432741 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49r5j\" (UniqueName: \"kubernetes.io/projected/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-kube-api-access-49r5j\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432958 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-log-ovn\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432998 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run-ovn\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.432973 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.433469 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-additional-scripts\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.434554 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-scripts\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.449628 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49r5j\" (UniqueName: \"kubernetes.io/projected/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-kube-api-access-49r5j\") pod \"ovn-controller-6frjn-config-prsrj\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.495385 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.580440 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.678256 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" event={"ID":"4167e405-5683-46fe-a8bb-f098143088ac","Type":"ContainerDied","Data":"96ba3517606911ed9d4ffb06c65882ff3336b7c843d6bedb24f1f5d934f072b6"} Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.678320 4990 scope.go:117] "RemoveContainer" containerID="d890669d3f2aec491fdaf9f0db83cd94c93ccc53c269f62091b0f31d9aad546d" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.678411 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8cc7fc4dc-fwm6p" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.706220 4990 scope.go:117] "RemoveContainer" containerID="8c903b2a2c9c457c79f1b01f606c38db2ff4273852ccd5c81a063266e32cb71d" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.738912 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-ovsdbserver-sb\") pod \"4167e405-5683-46fe-a8bb-f098143088ac\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.739033 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-dns-svc\") pod \"4167e405-5683-46fe-a8bb-f098143088ac\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.739096 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-config\") pod \"4167e405-5683-46fe-a8bb-f098143088ac\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.739122 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b98cg\" (UniqueName: \"kubernetes.io/projected/4167e405-5683-46fe-a8bb-f098143088ac-kube-api-access-b98cg\") pod \"4167e405-5683-46fe-a8bb-f098143088ac\" (UID: \"4167e405-5683-46fe-a8bb-f098143088ac\") " Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.743740 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4167e405-5683-46fe-a8bb-f098143088ac-kube-api-access-b98cg" (OuterVolumeSpecName: "kube-api-access-b98cg") pod "4167e405-5683-46fe-a8bb-f098143088ac" (UID: "4167e405-5683-46fe-a8bb-f098143088ac"). InnerVolumeSpecName "kube-api-access-b98cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.751987 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.784015 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-config" (OuterVolumeSpecName: "config") pod "4167e405-5683-46fe-a8bb-f098143088ac" (UID: "4167e405-5683-46fe-a8bb-f098143088ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.784886 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4167e405-5683-46fe-a8bb-f098143088ac" (UID: "4167e405-5683-46fe-a8bb-f098143088ac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.786302 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4167e405-5683-46fe-a8bb-f098143088ac" (UID: "4167e405-5683-46fe-a8bb-f098143088ac"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.841828 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.842117 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.842178 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4167e405-5683-46fe-a8bb-f098143088ac-config\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.842239 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b98cg\" (UniqueName: \"kubernetes.io/projected/4167e405-5683-46fe-a8bb-f098143088ac-kube-api-access-b98cg\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.914631 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 12:59:15 crc kubenswrapper[4990]: I1203 12:59:15.933853 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6frjn-config-prsrj"] Dec 03 12:59:16 crc kubenswrapper[4990]: I1203 12:59:16.011217 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-fwm6p"] Dec 03 12:59:16 crc kubenswrapper[4990]: I1203 12:59:16.017466 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8cc7fc4dc-fwm6p"] Dec 03 12:59:16 crc kubenswrapper[4990]: I1203 12:59:16.019129 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="81e14867-9f93-4820-ba0f-c6dbd12e31db" containerName="galera" probeResult="failure" output=< Dec 03 12:59:16 crc kubenswrapper[4990]: wsrep_local_state_comment (Joined) differs from Synced Dec 03 12:59:16 crc kubenswrapper[4990]: > Dec 03 12:59:16 crc kubenswrapper[4990]: I1203 12:59:16.276892 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4167e405-5683-46fe-a8bb-f098143088ac" path="/var/lib/kubelet/pods/4167e405-5683-46fe-a8bb-f098143088ac/volumes" Dec 03 12:59:16 crc kubenswrapper[4990]: I1203 12:59:16.277515 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8979791b-6642-4783-95a9-02a58fcaa400" path="/var/lib/kubelet/pods/8979791b-6642-4783-95a9-02a58fcaa400/volumes" Dec 03 12:59:16 crc kubenswrapper[4990]: I1203 12:59:16.686624 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-prsrj" event={"ID":"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff","Type":"ContainerStarted","Data":"777578b983385b82cfa8142b8da23a83fdb89b500cf0596498c22abaa0137626"} Dec 03 12:59:17 crc kubenswrapper[4990]: I1203 12:59:17.174203 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 12:59:17 crc kubenswrapper[4990]: I1203 12:59:17.174589 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 12:59:18 crc kubenswrapper[4990]: I1203 12:59:18.723266 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-prsrj" event={"ID":"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff","Type":"ContainerStarted","Data":"560c3677bbbafe6087addfdc243849349f0dedcf30e74d70b7c1ff66bc4250ca"} Dec 03 12:59:19 crc kubenswrapper[4990]: I1203 12:59:19.171331 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 12:59:23 crc kubenswrapper[4990]: I1203 12:59:23.108034 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 12:59:25 crc kubenswrapper[4990]: I1203 12:59:25.387744 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 03 12:59:25 crc kubenswrapper[4990]: I1203 12:59:25.750183 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 03 12:59:27 crc kubenswrapper[4990]: I1203 12:59:27.798843 4990 generic.go:334] "Generic (PLEG): container finished" podID="2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" containerID="560c3677bbbafe6087addfdc243849349f0dedcf30e74d70b7c1ff66bc4250ca" exitCode=0 Dec 03 12:59:27 crc kubenswrapper[4990]: I1203 12:59:27.798917 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-prsrj" event={"ID":"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff","Type":"ContainerDied","Data":"560c3677bbbafe6087addfdc243849349f0dedcf30e74d70b7c1ff66bc4250ca"} Dec 03 12:59:28 crc kubenswrapper[4990]: I1203 12:59:28.388657 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 12:59:28 crc kubenswrapper[4990]: I1203 12:59:28.472533 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="4fb6a200-beee-4b47-ac8d-e370e307070f" containerName="galera" probeResult="failure" output=< Dec 03 12:59:28 crc kubenswrapper[4990]: wsrep_local_state_comment (Joined) differs from Synced Dec 03 12:59:28 crc kubenswrapper[4990]: > Dec 03 12:59:28 crc kubenswrapper[4990]: I1203 12:59:28.808418 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gqdz2" event={"ID":"73f3c670-ab5f-4ab9-9278-9aa008c9a35b","Type":"ContainerStarted","Data":"779d21dd4e28d2a2230cf9160b45ad489d4113186d2f62322d1d03f20fa721f9"} Dec 03 12:59:28 crc kubenswrapper[4990]: I1203 12:59:28.835075 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-gqdz2" podStartSLOduration=2.216346097 podStartE2EDuration="45.835055435s" podCreationTimestamp="2025-12-03 12:58:43 +0000 UTC" firstStartedPulling="2025-12-03 12:58:44.137426231 +0000 UTC m=+1272.279337470" lastFinishedPulling="2025-12-03 12:59:27.756135569 +0000 UTC m=+1315.898046808" observedRunningTime="2025-12-03 12:59:28.831323086 +0000 UTC m=+1316.973234325" watchObservedRunningTime="2025-12-03 12:59:28.835055435 +0000 UTC m=+1316.976966664" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.169742 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283276 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-additional-scripts\") pod \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283724 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-scripts\") pod \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283760 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run-ovn\") pod \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283876 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49r5j\" (UniqueName: \"kubernetes.io/projected/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-kube-api-access-49r5j\") pod \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283932 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" (UID: "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283947 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run\") pod \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.283983 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run" (OuterVolumeSpecName: "var-run") pod "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" (UID: "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284049 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-log-ovn\") pod \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\" (UID: \"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff\") " Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284437 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" (UID: "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284508 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" (UID: "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284635 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-scripts" (OuterVolumeSpecName: "scripts") pod "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" (UID: "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284937 4990 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284961 4990 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284978 4990 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.284992 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.285001 4990 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.289271 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-kube-api-access-49r5j" (OuterVolumeSpecName: "kube-api-access-49r5j") pod "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" (UID: "2f68da4e-de1a-4e83-a3b0-bf0580fb64ff"). InnerVolumeSpecName "kube-api-access-49r5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.388730 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49r5j\" (UniqueName: \"kubernetes.io/projected/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff-kube-api-access-49r5j\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.817500 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6frjn-config-prsrj" event={"ID":"2f68da4e-de1a-4e83-a3b0-bf0580fb64ff","Type":"ContainerDied","Data":"777578b983385b82cfa8142b8da23a83fdb89b500cf0596498c22abaa0137626"} Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.817555 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="777578b983385b82cfa8142b8da23a83fdb89b500cf0596498c22abaa0137626" Dec 03 12:59:29 crc kubenswrapper[4990]: I1203 12:59:29.817583 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6frjn-config-prsrj" Dec 03 12:59:30 crc kubenswrapper[4990]: I1203 12:59:30.276090 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6frjn-config-prsrj"] Dec 03 12:59:30 crc kubenswrapper[4990]: I1203 12:59:30.285399 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6frjn-config-prsrj"] Dec 03 12:59:32 crc kubenswrapper[4990]: I1203 12:59:32.275008 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" path="/var/lib/kubelet/pods/2f68da4e-de1a-4e83-a3b0-bf0580fb64ff/volumes" Dec 03 12:59:33 crc kubenswrapper[4990]: I1203 12:59:33.286787 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 12:59:33 crc kubenswrapper[4990]: I1203 12:59:33.286854 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 12:59:35 crc kubenswrapper[4990]: I1203 12:59:35.386624 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 03 12:59:35 crc kubenswrapper[4990]: I1203 12:59:35.749539 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 03 12:59:37 crc kubenswrapper[4990]: I1203 12:59:37.260642 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.586227 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c9da-account-create-update-7zlqj"] Dec 03 12:59:38 crc kubenswrapper[4990]: E1203 12:59:38.586617 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4167e405-5683-46fe-a8bb-f098143088ac" containerName="dnsmasq-dns" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.586632 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4167e405-5683-46fe-a8bb-f098143088ac" containerName="dnsmasq-dns" Dec 03 12:59:38 crc kubenswrapper[4990]: E1203 12:59:38.586646 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4167e405-5683-46fe-a8bb-f098143088ac" containerName="init" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.586671 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4167e405-5683-46fe-a8bb-f098143088ac" containerName="init" Dec 03 12:59:38 crc kubenswrapper[4990]: E1203 12:59:38.586686 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" containerName="ovn-config" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.586693 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" containerName="ovn-config" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.586910 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4167e405-5683-46fe-a8bb-f098143088ac" containerName="dnsmasq-dns" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.586936 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f68da4e-de1a-4e83-a3b0-bf0580fb64ff" containerName="ovn-config" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.587580 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.590319 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.598485 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c9da-account-create-update-7zlqj"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.630324 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jdzvw"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.634333 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.642139 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jdzvw"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.678334 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-operator-scripts\") pod \"keystone-c9da-account-create-update-7zlqj\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.678546 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqn2b\" (UniqueName: \"kubernetes.io/projected/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-kube-api-access-rqn2b\") pod \"keystone-c9da-account-create-update-7zlqj\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.780012 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pmnq\" (UniqueName: \"kubernetes.io/projected/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-kube-api-access-2pmnq\") pod \"keystone-db-create-jdzvw\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.780237 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqn2b\" (UniqueName: \"kubernetes.io/projected/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-kube-api-access-rqn2b\") pod \"keystone-c9da-account-create-update-7zlqj\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.780288 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-operator-scripts\") pod \"keystone-db-create-jdzvw\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.780338 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-operator-scripts\") pod \"keystone-c9da-account-create-update-7zlqj\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.781462 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-operator-scripts\") pod \"keystone-c9da-account-create-update-7zlqj\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.799125 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqn2b\" (UniqueName: \"kubernetes.io/projected/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-kube-api-access-rqn2b\") pod \"keystone-c9da-account-create-update-7zlqj\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.828021 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-6fs42"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.829852 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6fs42" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.850429 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6fs42"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.881712 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-operator-scripts\") pod \"keystone-db-create-jdzvw\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.881770 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pmnq\" (UniqueName: \"kubernetes.io/projected/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-kube-api-access-2pmnq\") pod \"keystone-db-create-jdzvw\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.882628 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-operator-scripts\") pod \"keystone-db-create-jdzvw\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.909323 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pmnq\" (UniqueName: \"kubernetes.io/projected/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-kube-api-access-2pmnq\") pod \"keystone-db-create-jdzvw\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.911060 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.953620 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.967973 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3d39-account-create-update-vwp7z"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.969413 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.975869 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.982654 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3d39-account-create-update-vwp7z"] Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.983513 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03cec11-67cd-483a-929e-003f234f3c6d-operator-scripts\") pod \"placement-db-create-6fs42\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " pod="openstack/placement-db-create-6fs42" Dec 03 12:59:38 crc kubenswrapper[4990]: I1203 12:59:38.983736 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jksx8\" (UniqueName: \"kubernetes.io/projected/a03cec11-67cd-483a-929e-003f234f3c6d-kube-api-access-jksx8\") pod \"placement-db-create-6fs42\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " pod="openstack/placement-db-create-6fs42" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.085439 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9zz4\" (UniqueName: \"kubernetes.io/projected/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-kube-api-access-b9zz4\") pod \"placement-3d39-account-create-update-vwp7z\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.085954 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-operator-scripts\") pod \"placement-3d39-account-create-update-vwp7z\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.086041 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jksx8\" (UniqueName: \"kubernetes.io/projected/a03cec11-67cd-483a-929e-003f234f3c6d-kube-api-access-jksx8\") pod \"placement-db-create-6fs42\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " pod="openstack/placement-db-create-6fs42" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.086113 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03cec11-67cd-483a-929e-003f234f3c6d-operator-scripts\") pod \"placement-db-create-6fs42\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " pod="openstack/placement-db-create-6fs42" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.087108 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03cec11-67cd-483a-929e-003f234f3c6d-operator-scripts\") pod \"placement-db-create-6fs42\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " pod="openstack/placement-db-create-6fs42" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.114204 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jksx8\" (UniqueName: \"kubernetes.io/projected/a03cec11-67cd-483a-929e-003f234f3c6d-kube-api-access-jksx8\") pod \"placement-db-create-6fs42\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " pod="openstack/placement-db-create-6fs42" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.141365 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-mtcnj"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.142478 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.173824 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mtcnj"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.189856 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9zz4\" (UniqueName: \"kubernetes.io/projected/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-kube-api-access-b9zz4\") pod \"placement-3d39-account-create-update-vwp7z\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.189951 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-operator-scripts\") pod \"placement-3d39-account-create-update-vwp7z\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.194045 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-operator-scripts\") pod \"placement-3d39-account-create-update-vwp7z\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.203537 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6fs42" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.212851 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9zz4\" (UniqueName: \"kubernetes.io/projected/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-kube-api-access-b9zz4\") pod \"placement-3d39-account-create-update-vwp7z\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.261109 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-dd77-account-create-update-bl6tk"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.262419 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.265168 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.290528 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-dd77-account-create-update-bl6tk"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.291328 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00bf24a-fcbf-4b46-b502-3c246b290851-operator-scripts\") pod \"glance-db-create-mtcnj\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.291495 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvx7g\" (UniqueName: \"kubernetes.io/projected/e00bf24a-fcbf-4b46-b502-3c246b290851-kube-api-access-rvx7g\") pod \"glance-db-create-mtcnj\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.393873 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00bf24a-fcbf-4b46-b502-3c246b290851-operator-scripts\") pod \"glance-db-create-mtcnj\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.394824 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140fdc8c-c0ba-4761-a398-cfed9c1179b4-operator-scripts\") pod \"glance-dd77-account-create-update-bl6tk\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.394947 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvx7g\" (UniqueName: \"kubernetes.io/projected/e00bf24a-fcbf-4b46-b502-3c246b290851-kube-api-access-rvx7g\") pod \"glance-db-create-mtcnj\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.395031 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrz9t\" (UniqueName: \"kubernetes.io/projected/140fdc8c-c0ba-4761-a398-cfed9c1179b4-kube-api-access-jrz9t\") pod \"glance-dd77-account-create-update-bl6tk\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.395147 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00bf24a-fcbf-4b46-b502-3c246b290851-operator-scripts\") pod \"glance-db-create-mtcnj\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.411946 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.418780 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvx7g\" (UniqueName: \"kubernetes.io/projected/e00bf24a-fcbf-4b46-b502-3c246b290851-kube-api-access-rvx7g\") pod \"glance-db-create-mtcnj\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.468662 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.496839 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrz9t\" (UniqueName: \"kubernetes.io/projected/140fdc8c-c0ba-4761-a398-cfed9c1179b4-kube-api-access-jrz9t\") pod \"glance-dd77-account-create-update-bl6tk\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.496985 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140fdc8c-c0ba-4761-a398-cfed9c1179b4-operator-scripts\") pod \"glance-dd77-account-create-update-bl6tk\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.498008 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140fdc8c-c0ba-4761-a398-cfed9c1179b4-operator-scripts\") pod \"glance-dd77-account-create-update-bl6tk\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.518590 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrz9t\" (UniqueName: \"kubernetes.io/projected/140fdc8c-c0ba-4761-a398-cfed9c1179b4-kube-api-access-jrz9t\") pod \"glance-dd77-account-create-update-bl6tk\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.557852 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c9da-account-create-update-7zlqj"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.591127 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.902748 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6fs42"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.917250 4990 generic.go:334] "Generic (PLEG): container finished" podID="73f3c670-ab5f-4ab9-9278-9aa008c9a35b" containerID="779d21dd4e28d2a2230cf9160b45ad489d4113186d2f62322d1d03f20fa721f9" exitCode=0 Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.917329 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gqdz2" event={"ID":"73f3c670-ab5f-4ab9-9278-9aa008c9a35b","Type":"ContainerDied","Data":"779d21dd4e28d2a2230cf9160b45ad489d4113186d2f62322d1d03f20fa721f9"} Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.929843 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3d39-account-create-update-vwp7z"] Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.934744 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9da-account-create-update-7zlqj" event={"ID":"743c06a8-5f6c-4a31-9a4b-08f15170d3b0","Type":"ContainerStarted","Data":"68dc5b375e77f873c7d199419aeafde4d1e1658aa9215bd3829e0636e19f4cc4"} Dec 03 12:59:39 crc kubenswrapper[4990]: I1203 12:59:39.973574 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jdzvw"] Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.027377 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mtcnj"] Dec 03 12:59:40 crc kubenswrapper[4990]: W1203 12:59:40.036922 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode00bf24a_fcbf_4b46_b502_3c246b290851.slice/crio-e86fb5be030eec26c1dbab92d6c4d56fb0724abfcd1fdea2a9a4758cb9e1d7b3 WatchSource:0}: Error finding container e86fb5be030eec26c1dbab92d6c4d56fb0724abfcd1fdea2a9a4758cb9e1d7b3: Status 404 returned error can't find the container with id e86fb5be030eec26c1dbab92d6c4d56fb0724abfcd1fdea2a9a4758cb9e1d7b3 Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.299257 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-dd77-account-create-update-bl6tk"] Dec 03 12:59:40 crc kubenswrapper[4990]: W1203 12:59:40.302336 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod140fdc8c_c0ba_4761_a398_cfed9c1179b4.slice/crio-fb2418d57e62ab464769b1cdc97f58e04b7bf5268f586a605c916648e9a60bb7 WatchSource:0}: Error finding container fb2418d57e62ab464769b1cdc97f58e04b7bf5268f586a605c916648e9a60bb7: Status 404 returned error can't find the container with id fb2418d57e62ab464769b1cdc97f58e04b7bf5268f586a605c916648e9a60bb7 Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.943962 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d39-account-create-update-vwp7z" event={"ID":"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9","Type":"ContainerStarted","Data":"3464588fc087417713056f848bac5c229c60657762b721327e4d3682fbc59af6"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.945193 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d39-account-create-update-vwp7z" event={"ID":"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9","Type":"ContainerStarted","Data":"ea40fe3fa4ac49bf3e2194cce86876ba2f2947cd15ac5aadb80fca7a13b5f57a"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.946577 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dd77-account-create-update-bl6tk" event={"ID":"140fdc8c-c0ba-4761-a398-cfed9c1179b4","Type":"ContainerStarted","Data":"9575d22f08298c6a6ad0a7c692b075460e1f831b6a312b85fa105310096cc9a9"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.946632 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dd77-account-create-update-bl6tk" event={"ID":"140fdc8c-c0ba-4761-a398-cfed9c1179b4","Type":"ContainerStarted","Data":"fb2418d57e62ab464769b1cdc97f58e04b7bf5268f586a605c916648e9a60bb7"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.949558 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mtcnj" event={"ID":"e00bf24a-fcbf-4b46-b502-3c246b290851","Type":"ContainerStarted","Data":"0325b42ca6ed2d7abce4a6799e178f4a7dc9f71a3fd1c5e630a8af28758f6f80"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.949697 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mtcnj" event={"ID":"e00bf24a-fcbf-4b46-b502-3c246b290851","Type":"ContainerStarted","Data":"e86fb5be030eec26c1dbab92d6c4d56fb0724abfcd1fdea2a9a4758cb9e1d7b3"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.951033 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6fs42" event={"ID":"a03cec11-67cd-483a-929e-003f234f3c6d","Type":"ContainerStarted","Data":"6cdf946732ae067ce4ed96be4c91799ef9763df56be8b2395aae5e057086b982"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.951084 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6fs42" event={"ID":"a03cec11-67cd-483a-929e-003f234f3c6d","Type":"ContainerStarted","Data":"f856df37eb8e5015701d403676f490e6ed8c905f97c89683c45eff481e7005f6"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.956144 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9da-account-create-update-7zlqj" event={"ID":"743c06a8-5f6c-4a31-9a4b-08f15170d3b0","Type":"ContainerStarted","Data":"f454a96e8441abaca7cd2ff490cb86afc362ed9140888a2d14eb1f0cf55a7790"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.964352 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdzvw" event={"ID":"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0","Type":"ContainerStarted","Data":"381554b619f1bfc51a2f37272a046996b23f7b1d933e399fc35229629e1f49a8"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.964424 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdzvw" event={"ID":"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0","Type":"ContainerStarted","Data":"77f72e91e997627d8d26867bddc93e8b932d9bb57ab79a1ef03afcf6e8d70b3b"} Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.969308 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-3d39-account-create-update-vwp7z" podStartSLOduration=2.969177599 podStartE2EDuration="2.969177599s" podCreationTimestamp="2025-12-03 12:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:40.96009237 +0000 UTC m=+1329.102003599" watchObservedRunningTime="2025-12-03 12:59:40.969177599 +0000 UTC m=+1329.111088828" Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.996994 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-dd77-account-create-update-bl6tk" podStartSLOduration=1.996975252 podStartE2EDuration="1.996975252s" podCreationTimestamp="2025-12-03 12:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:40.981716689 +0000 UTC m=+1329.123627918" watchObservedRunningTime="2025-12-03 12:59:40.996975252 +0000 UTC m=+1329.138886481" Dec 03 12:59:40 crc kubenswrapper[4990]: I1203 12:59:40.997660 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-6fs42" podStartSLOduration=2.99765362 podStartE2EDuration="2.99765362s" podCreationTimestamp="2025-12-03 12:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:40.994282781 +0000 UTC m=+1329.136194020" watchObservedRunningTime="2025-12-03 12:59:40.99765362 +0000 UTC m=+1329.139564849" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.011929 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c9da-account-create-update-7zlqj" podStartSLOduration=3.011904856 podStartE2EDuration="3.011904856s" podCreationTimestamp="2025-12-03 12:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:41.007057848 +0000 UTC m=+1329.148969097" watchObservedRunningTime="2025-12-03 12:59:41.011904856 +0000 UTC m=+1329.153816085" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.032832 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-mtcnj" podStartSLOduration=2.032812597 podStartE2EDuration="2.032812597s" podCreationTimestamp="2025-12-03 12:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:41.031666577 +0000 UTC m=+1329.173577806" watchObservedRunningTime="2025-12-03 12:59:41.032812597 +0000 UTC m=+1329.174723826" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.093200 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-jdzvw" podStartSLOduration=3.093180309 podStartE2EDuration="3.093180309s" podCreationTimestamp="2025-12-03 12:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 12:59:41.090620701 +0000 UTC m=+1329.232531930" watchObservedRunningTime="2025-12-03 12:59:41.093180309 +0000 UTC m=+1329.235091538" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.534893 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640489 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-etc-swift\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640564 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-swiftconf\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640603 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-dispersionconf\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640628 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l62kc\" (UniqueName: \"kubernetes.io/projected/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-kube-api-access-l62kc\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640705 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-ring-data-devices\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640738 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-combined-ca-bundle\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.640817 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-scripts\") pod \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\" (UID: \"73f3c670-ab5f-4ab9-9278-9aa008c9a35b\") " Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.641556 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.642473 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.649746 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.652232 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-kube-api-access-l62kc" (OuterVolumeSpecName: "kube-api-access-l62kc") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "kube-api-access-l62kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.667002 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-scripts" (OuterVolumeSpecName: "scripts") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.670750 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.675150 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "73f3c670-ab5f-4ab9-9278-9aa008c9a35b" (UID: "73f3c670-ab5f-4ab9-9278-9aa008c9a35b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742600 4990 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742651 4990 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742662 4990 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742676 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l62kc\" (UniqueName: \"kubernetes.io/projected/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-kube-api-access-l62kc\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742692 4990 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742703 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.742717 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f3c670-ab5f-4ab9-9278-9aa008c9a35b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.976151 4990 generic.go:334] "Generic (PLEG): container finished" podID="e00bf24a-fcbf-4b46-b502-3c246b290851" containerID="0325b42ca6ed2d7abce4a6799e178f4a7dc9f71a3fd1c5e630a8af28758f6f80" exitCode=0 Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.976244 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mtcnj" event={"ID":"e00bf24a-fcbf-4b46-b502-3c246b290851","Type":"ContainerDied","Data":"0325b42ca6ed2d7abce4a6799e178f4a7dc9f71a3fd1c5e630a8af28758f6f80"} Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.977798 4990 generic.go:334] "Generic (PLEG): container finished" podID="a03cec11-67cd-483a-929e-003f234f3c6d" containerID="6cdf946732ae067ce4ed96be4c91799ef9763df56be8b2395aae5e057086b982" exitCode=0 Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.977853 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6fs42" event={"ID":"a03cec11-67cd-483a-929e-003f234f3c6d","Type":"ContainerDied","Data":"6cdf946732ae067ce4ed96be4c91799ef9763df56be8b2395aae5e057086b982"} Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.979744 4990 generic.go:334] "Generic (PLEG): container finished" podID="743c06a8-5f6c-4a31-9a4b-08f15170d3b0" containerID="f454a96e8441abaca7cd2ff490cb86afc362ed9140888a2d14eb1f0cf55a7790" exitCode=0 Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.979882 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9da-account-create-update-7zlqj" event={"ID":"743c06a8-5f6c-4a31-9a4b-08f15170d3b0","Type":"ContainerDied","Data":"f454a96e8441abaca7cd2ff490cb86afc362ed9140888a2d14eb1f0cf55a7790"} Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.983704 4990 generic.go:334] "Generic (PLEG): container finished" podID="2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" containerID="381554b619f1bfc51a2f37272a046996b23f7b1d933e399fc35229629e1f49a8" exitCode=0 Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.983795 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdzvw" event={"ID":"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0","Type":"ContainerDied","Data":"381554b619f1bfc51a2f37272a046996b23f7b1d933e399fc35229629e1f49a8"} Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.986123 4990 generic.go:334] "Generic (PLEG): container finished" podID="5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" containerID="3464588fc087417713056f848bac5c229c60657762b721327e4d3682fbc59af6" exitCode=0 Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.986229 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d39-account-create-update-vwp7z" event={"ID":"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9","Type":"ContainerDied","Data":"3464588fc087417713056f848bac5c229c60657762b721327e4d3682fbc59af6"} Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.987873 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gqdz2" event={"ID":"73f3c670-ab5f-4ab9-9278-9aa008c9a35b","Type":"ContainerDied","Data":"67c00efb09ca140e387ad553554db9a449bba770f221723a550dc65ed14162f0"} Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.987907 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67c00efb09ca140e387ad553554db9a449bba770f221723a550dc65ed14162f0" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.987963 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gqdz2" Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.990724 4990 generic.go:334] "Generic (PLEG): container finished" podID="140fdc8c-c0ba-4761-a398-cfed9c1179b4" containerID="9575d22f08298c6a6ad0a7c692b075460e1f831b6a312b85fa105310096cc9a9" exitCode=0 Dec 03 12:59:41 crc kubenswrapper[4990]: I1203 12:59:41.990793 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dd77-account-create-update-bl6tk" event={"ID":"140fdc8c-c0ba-4761-a398-cfed9c1179b4","Type":"ContainerDied","Data":"9575d22f08298c6a6ad0a7c692b075460e1f831b6a312b85fa105310096cc9a9"} Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.400562 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.473834 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140fdc8c-c0ba-4761-a398-cfed9c1179b4-operator-scripts\") pod \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.473978 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrz9t\" (UniqueName: \"kubernetes.io/projected/140fdc8c-c0ba-4761-a398-cfed9c1179b4-kube-api-access-jrz9t\") pod \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\" (UID: \"140fdc8c-c0ba-4761-a398-cfed9c1179b4\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.475269 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140fdc8c-c0ba-4761-a398-cfed9c1179b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "140fdc8c-c0ba-4761-a398-cfed9c1179b4" (UID: "140fdc8c-c0ba-4761-a398-cfed9c1179b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.481541 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140fdc8c-c0ba-4761-a398-cfed9c1179b4-kube-api-access-jrz9t" (OuterVolumeSpecName: "kube-api-access-jrz9t") pod "140fdc8c-c0ba-4761-a398-cfed9c1179b4" (UID: "140fdc8c-c0ba-4761-a398-cfed9c1179b4"). InnerVolumeSpecName "kube-api-access-jrz9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.576471 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140fdc8c-c0ba-4761-a398-cfed9c1179b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.576523 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrz9t\" (UniqueName: \"kubernetes.io/projected/140fdc8c-c0ba-4761-a398-cfed9c1179b4-kube-api-access-jrz9t\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.592383 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.597859 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6fs42" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.620467 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.622111 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.645782 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677104 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03cec11-67cd-483a-929e-003f234f3c6d-operator-scripts\") pod \"a03cec11-67cd-483a-929e-003f234f3c6d\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677161 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-operator-scripts\") pod \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677228 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jksx8\" (UniqueName: \"kubernetes.io/projected/a03cec11-67cd-483a-929e-003f234f3c6d-kube-api-access-jksx8\") pod \"a03cec11-67cd-483a-929e-003f234f3c6d\" (UID: \"a03cec11-67cd-483a-929e-003f234f3c6d\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677293 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00bf24a-fcbf-4b46-b502-3c246b290851-operator-scripts\") pod \"e00bf24a-fcbf-4b46-b502-3c246b290851\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677321 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-operator-scripts\") pod \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677387 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvx7g\" (UniqueName: \"kubernetes.io/projected/e00bf24a-fcbf-4b46-b502-3c246b290851-kube-api-access-rvx7g\") pod \"e00bf24a-fcbf-4b46-b502-3c246b290851\" (UID: \"e00bf24a-fcbf-4b46-b502-3c246b290851\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677481 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pmnq\" (UniqueName: \"kubernetes.io/projected/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-kube-api-access-2pmnq\") pod \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\" (UID: \"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.677532 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqn2b\" (UniqueName: \"kubernetes.io/projected/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-kube-api-access-rqn2b\") pod \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\" (UID: \"743c06a8-5f6c-4a31-9a4b-08f15170d3b0\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.679894 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e00bf24a-fcbf-4b46-b502-3c246b290851-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e00bf24a-fcbf-4b46-b502-3c246b290851" (UID: "e00bf24a-fcbf-4b46-b502-3c246b290851"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.681312 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" (UID: "2cd3baf6-2af2-41e9-a3c4-a31fe32588f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.681361 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "743c06a8-5f6c-4a31-9a4b-08f15170d3b0" (UID: "743c06a8-5f6c-4a31-9a4b-08f15170d3b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.682756 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03cec11-67cd-483a-929e-003f234f3c6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a03cec11-67cd-483a-929e-003f234f3c6d" (UID: "a03cec11-67cd-483a-929e-003f234f3c6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.683935 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-kube-api-access-rqn2b" (OuterVolumeSpecName: "kube-api-access-rqn2b") pod "743c06a8-5f6c-4a31-9a4b-08f15170d3b0" (UID: "743c06a8-5f6c-4a31-9a4b-08f15170d3b0"). InnerVolumeSpecName "kube-api-access-rqn2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.685627 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-kube-api-access-2pmnq" (OuterVolumeSpecName: "kube-api-access-2pmnq") pod "2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" (UID: "2cd3baf6-2af2-41e9-a3c4-a31fe32588f0"). InnerVolumeSpecName "kube-api-access-2pmnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.686004 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e00bf24a-fcbf-4b46-b502-3c246b290851-kube-api-access-rvx7g" (OuterVolumeSpecName: "kube-api-access-rvx7g") pod "e00bf24a-fcbf-4b46-b502-3c246b290851" (UID: "e00bf24a-fcbf-4b46-b502-3c246b290851"). InnerVolumeSpecName "kube-api-access-rvx7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.688004 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a03cec11-67cd-483a-929e-003f234f3c6d-kube-api-access-jksx8" (OuterVolumeSpecName: "kube-api-access-jksx8") pod "a03cec11-67cd-483a-929e-003f234f3c6d" (UID: "a03cec11-67cd-483a-929e-003f234f3c6d"). InnerVolumeSpecName "kube-api-access-jksx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780238 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-operator-scripts\") pod \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780378 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9zz4\" (UniqueName: \"kubernetes.io/projected/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-kube-api-access-b9zz4\") pod \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\" (UID: \"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9\") " Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780852 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00bf24a-fcbf-4b46-b502-3c246b290851-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780868 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780876 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvx7g\" (UniqueName: \"kubernetes.io/projected/e00bf24a-fcbf-4b46-b502-3c246b290851-kube-api-access-rvx7g\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780886 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pmnq\" (UniqueName: \"kubernetes.io/projected/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0-kube-api-access-2pmnq\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780895 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqn2b\" (UniqueName: \"kubernetes.io/projected/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-kube-api-access-rqn2b\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780903 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03cec11-67cd-483a-929e-003f234f3c6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780911 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/743c06a8-5f6c-4a31-9a4b-08f15170d3b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.780920 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jksx8\" (UniqueName: \"kubernetes.io/projected/a03cec11-67cd-483a-929e-003f234f3c6d-kube-api-access-jksx8\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.781641 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" (UID: "5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.787738 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-kube-api-access-b9zz4" (OuterVolumeSpecName: "kube-api-access-b9zz4") pod "5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" (UID: "5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9"). InnerVolumeSpecName "kube-api-access-b9zz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.882590 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9zz4\" (UniqueName: \"kubernetes.io/projected/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-kube-api-access-b9zz4\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:43 crc kubenswrapper[4990]: I1203 12:59:43.882640 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.011494 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jdzvw" event={"ID":"2cd3baf6-2af2-41e9-a3c4-a31fe32588f0","Type":"ContainerDied","Data":"77f72e91e997627d8d26867bddc93e8b932d9bb57ab79a1ef03afcf6e8d70b3b"} Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.011534 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77f72e91e997627d8d26867bddc93e8b932d9bb57ab79a1ef03afcf6e8d70b3b" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.011600 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jdzvw" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.024348 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d39-account-create-update-vwp7z" event={"ID":"5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9","Type":"ContainerDied","Data":"ea40fe3fa4ac49bf3e2194cce86876ba2f2947cd15ac5aadb80fca7a13b5f57a"} Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.024400 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea40fe3fa4ac49bf3e2194cce86876ba2f2947cd15ac5aadb80fca7a13b5f57a" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.024579 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d39-account-create-update-vwp7z" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.027267 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dd77-account-create-update-bl6tk" event={"ID":"140fdc8c-c0ba-4761-a398-cfed9c1179b4","Type":"ContainerDied","Data":"fb2418d57e62ab464769b1cdc97f58e04b7bf5268f586a605c916648e9a60bb7"} Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.027322 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb2418d57e62ab464769b1cdc97f58e04b7bf5268f586a605c916648e9a60bb7" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.027411 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dd77-account-create-update-bl6tk" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.044877 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mtcnj" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.045532 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mtcnj" event={"ID":"e00bf24a-fcbf-4b46-b502-3c246b290851","Type":"ContainerDied","Data":"e86fb5be030eec26c1dbab92d6c4d56fb0724abfcd1fdea2a9a4758cb9e1d7b3"} Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.045712 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e86fb5be030eec26c1dbab92d6c4d56fb0724abfcd1fdea2a9a4758cb9e1d7b3" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.047466 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6fs42" event={"ID":"a03cec11-67cd-483a-929e-003f234f3c6d","Type":"ContainerDied","Data":"f856df37eb8e5015701d403676f490e6ed8c905f97c89683c45eff481e7005f6"} Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.047502 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f856df37eb8e5015701d403676f490e6ed8c905f97c89683c45eff481e7005f6" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.047567 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6fs42" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.053715 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9da-account-create-update-7zlqj" event={"ID":"743c06a8-5f6c-4a31-9a4b-08f15170d3b0","Type":"ContainerDied","Data":"68dc5b375e77f873c7d199419aeafde4d1e1658aa9215bd3829e0636e19f4cc4"} Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.053776 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68dc5b375e77f873c7d199419aeafde4d1e1658aa9215bd3829e0636e19f4cc4" Dec 03 12:59:44 crc kubenswrapper[4990]: I1203 12:59:44.053775 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9da-account-create-update-7zlqj" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.389661 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.752842 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815202 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rx7m6"] Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815736 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f3c670-ab5f-4ab9-9278-9aa008c9a35b" containerName="swift-ring-rebalance" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815760 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f3c670-ab5f-4ab9-9278-9aa008c9a35b" containerName="swift-ring-rebalance" Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815776 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03cec11-67cd-483a-929e-003f234f3c6d" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815783 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03cec11-67cd-483a-929e-003f234f3c6d" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815792 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743c06a8-5f6c-4a31-9a4b-08f15170d3b0" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815800 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="743c06a8-5f6c-4a31-9a4b-08f15170d3b0" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815808 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00bf24a-fcbf-4b46-b502-3c246b290851" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815814 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00bf24a-fcbf-4b46-b502-3c246b290851" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815832 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815838 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815851 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815857 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: E1203 12:59:45.815866 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="140fdc8c-c0ba-4761-a398-cfed9c1179b4" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.815872 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="140fdc8c-c0ba-4761-a398-cfed9c1179b4" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816056 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="743c06a8-5f6c-4a31-9a4b-08f15170d3b0" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816065 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816086 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f3c670-ab5f-4ab9-9278-9aa008c9a35b" containerName="swift-ring-rebalance" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816099 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="140fdc8c-c0ba-4761-a398-cfed9c1179b4" containerName="mariadb-account-create-update" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816110 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816118 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="e00bf24a-fcbf-4b46-b502-3c246b290851" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816127 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="a03cec11-67cd-483a-929e-003f234f3c6d" containerName="mariadb-database-create" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.816772 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.823612 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d8c5-account-create-update-qhz9j"] Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.824912 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.827370 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.864020 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d8c5-account-create-update-qhz9j"] Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.896527 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rx7m6"] Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.923922 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-operator-scripts\") pod \"barbican-d8c5-account-create-update-qhz9j\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.924016 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z4kr\" (UniqueName: \"kubernetes.io/projected/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-kube-api-access-8z4kr\") pod \"barbican-d8c5-account-create-update-qhz9j\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.924132 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c5047b8-3306-40b5-bbaf-a30c6286662b-operator-scripts\") pod \"barbican-db-create-rx7m6\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.924272 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsl8d\" (UniqueName: \"kubernetes.io/projected/1c5047b8-3306-40b5-bbaf-a30c6286662b-kube-api-access-fsl8d\") pod \"barbican-db-create-rx7m6\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.994353 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-kkvhj"] Dec 03 12:59:45 crc kubenswrapper[4990]: I1203 12:59:45.995776 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.005203 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5066-account-create-update-w7bck"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.007117 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.008909 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.029035 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kkvhj"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.029886 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95aafce7-81a5-45a4-aa39-82431c05b665-operator-scripts\") pod \"cinder-db-create-kkvhj\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.029938 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-operator-scripts\") pod \"barbican-d8c5-account-create-update-qhz9j\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.029960 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3879fb-c88c-4c57-9620-d620812aa5cf-operator-scripts\") pod \"cinder-5066-account-create-update-w7bck\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.029982 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp4rl\" (UniqueName: \"kubernetes.io/projected/2e3879fb-c88c-4c57-9620-d620812aa5cf-kube-api-access-rp4rl\") pod \"cinder-5066-account-create-update-w7bck\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.030019 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z4kr\" (UniqueName: \"kubernetes.io/projected/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-kube-api-access-8z4kr\") pod \"barbican-d8c5-account-create-update-qhz9j\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.030047 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c5047b8-3306-40b5-bbaf-a30c6286662b-operator-scripts\") pod \"barbican-db-create-rx7m6\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.030094 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn857\" (UniqueName: \"kubernetes.io/projected/95aafce7-81a5-45a4-aa39-82431c05b665-kube-api-access-pn857\") pod \"cinder-db-create-kkvhj\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.030125 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsl8d\" (UniqueName: \"kubernetes.io/projected/1c5047b8-3306-40b5-bbaf-a30c6286662b-kube-api-access-fsl8d\") pod \"barbican-db-create-rx7m6\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.031112 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-operator-scripts\") pod \"barbican-d8c5-account-create-update-qhz9j\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.031666 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c5047b8-3306-40b5-bbaf-a30c6286662b-operator-scripts\") pod \"barbican-db-create-rx7m6\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.037525 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5066-account-create-update-w7bck"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.057627 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsl8d\" (UniqueName: \"kubernetes.io/projected/1c5047b8-3306-40b5-bbaf-a30c6286662b-kube-api-access-fsl8d\") pod \"barbican-db-create-rx7m6\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.061280 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z4kr\" (UniqueName: \"kubernetes.io/projected/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-kube-api-access-8z4kr\") pod \"barbican-d8c5-account-create-update-qhz9j\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.103085 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-gjl5k"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.104081 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.117300 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-gjl5k"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.132725 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95aafce7-81a5-45a4-aa39-82431c05b665-operator-scripts\") pod \"cinder-db-create-kkvhj\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.132777 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3879fb-c88c-4c57-9620-d620812aa5cf-operator-scripts\") pod \"cinder-5066-account-create-update-w7bck\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.132805 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp4rl\" (UniqueName: \"kubernetes.io/projected/2e3879fb-c88c-4c57-9620-d620812aa5cf-kube-api-access-rp4rl\") pod \"cinder-5066-account-create-update-w7bck\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.132895 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txn2l\" (UniqueName: \"kubernetes.io/projected/45d0ab64-a5d0-4e19-8b39-d38906823542-kube-api-access-txn2l\") pod \"neutron-db-create-gjl5k\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.132917 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn857\" (UniqueName: \"kubernetes.io/projected/95aafce7-81a5-45a4-aa39-82431c05b665-kube-api-access-pn857\") pod \"cinder-db-create-kkvhj\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.132944 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d0ab64-a5d0-4e19-8b39-d38906823542-operator-scripts\") pod \"neutron-db-create-gjl5k\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.133617 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95aafce7-81a5-45a4-aa39-82431c05b665-operator-scripts\") pod \"cinder-db-create-kkvhj\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.134053 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3879fb-c88c-4c57-9620-d620812aa5cf-operator-scripts\") pod \"cinder-5066-account-create-update-w7bck\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.143942 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.152847 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp4rl\" (UniqueName: \"kubernetes.io/projected/2e3879fb-c88c-4c57-9620-d620812aa5cf-kube-api-access-rp4rl\") pod \"cinder-5066-account-create-update-w7bck\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.153556 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.153850 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn857\" (UniqueName: \"kubernetes.io/projected/95aafce7-81a5-45a4-aa39-82431c05b665-kube-api-access-pn857\") pod \"cinder-db-create-kkvhj\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.236002 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txn2l\" (UniqueName: \"kubernetes.io/projected/45d0ab64-a5d0-4e19-8b39-d38906823542-kube-api-access-txn2l\") pod \"neutron-db-create-gjl5k\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.236079 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d0ab64-a5d0-4e19-8b39-d38906823542-operator-scripts\") pod \"neutron-db-create-gjl5k\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.237194 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d0ab64-a5d0-4e19-8b39-d38906823542-operator-scripts\") pod \"neutron-db-create-gjl5k\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.258225 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txn2l\" (UniqueName: \"kubernetes.io/projected/45d0ab64-a5d0-4e19-8b39-d38906823542-kube-api-access-txn2l\") pod \"neutron-db-create-gjl5k\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.259858 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.313939 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.326643 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-29e5-account-create-update-t945b"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.327781 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.331747 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.337038 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.354320 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-29e5-account-create-update-t945b"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.439165 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhxg2\" (UniqueName: \"kubernetes.io/projected/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-kube-api-access-nhxg2\") pod \"neutron-29e5-account-create-update-t945b\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.440055 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-operator-scripts\") pod \"neutron-29e5-account-create-update-t945b\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.520397 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d8c5-account-create-update-qhz9j"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.542557 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhxg2\" (UniqueName: \"kubernetes.io/projected/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-kube-api-access-nhxg2\") pod \"neutron-29e5-account-create-update-t945b\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.542633 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-operator-scripts\") pod \"neutron-29e5-account-create-update-t945b\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.543495 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-operator-scripts\") pod \"neutron-29e5-account-create-update-t945b\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.567827 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhxg2\" (UniqueName: \"kubernetes.io/projected/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-kube-api-access-nhxg2\") pod \"neutron-29e5-account-create-update-t945b\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.655185 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rx7m6"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.670789 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.749663 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.755530 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7d434c5-0459-4d5a-a401-4a0c4b82a553-etc-swift\") pod \"swift-storage-0\" (UID: \"f7d434c5-0459-4d5a-a401-4a0c4b82a553\") " pod="openstack/swift-storage-0" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.835753 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5066-account-create-update-w7bck"] Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.909321 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-gjl5k"] Dec 03 12:59:46 crc kubenswrapper[4990]: W1203 12:59:46.935833 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d0ab64_a5d0_4e19_8b39_d38906823542.slice/crio-1c9a51f8ab27d1559bc31c0effe779cebdfb724dfe08ea8fe6f9bafc0bcb84f3 WatchSource:0}: Error finding container 1c9a51f8ab27d1559bc31c0effe779cebdfb724dfe08ea8fe6f9bafc0bcb84f3: Status 404 returned error can't find the container with id 1c9a51f8ab27d1559bc31c0effe779cebdfb724dfe08ea8fe6f9bafc0bcb84f3 Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.948679 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 12:59:46 crc kubenswrapper[4990]: I1203 12:59:46.978529 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kkvhj"] Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.081181 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5066-account-create-update-w7bck" event={"ID":"2e3879fb-c88c-4c57-9620-d620812aa5cf","Type":"ContainerStarted","Data":"4aff73f8877276a0f8632941b829e74bb349868dbfdb98e8354b539d210d2045"} Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.083867 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d8c5-account-create-update-qhz9j" event={"ID":"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7","Type":"ContainerStarted","Data":"4285c328bd0d853460ef5ada0b3e625a920d091d4aa543f4a6ec6578ea364c68"} Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.085086 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjl5k" event={"ID":"45d0ab64-a5d0-4e19-8b39-d38906823542","Type":"ContainerStarted","Data":"1c9a51f8ab27d1559bc31c0effe779cebdfb724dfe08ea8fe6f9bafc0bcb84f3"} Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.086119 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rx7m6" event={"ID":"1c5047b8-3306-40b5-bbaf-a30c6286662b","Type":"ContainerStarted","Data":"12fd6adad3a25ced5d0c4c136532e4c306b56625e50e8bae2da34d4627866adb"} Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.089084 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kkvhj" event={"ID":"95aafce7-81a5-45a4-aa39-82431c05b665","Type":"ContainerStarted","Data":"7719b4a1826e3ef3bdc324a3561354073c6fa7ada94eaf7caf343510a751b1e5"} Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.228319 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-29e5-account-create-update-t945b"] Dec 03 12:59:47 crc kubenswrapper[4990]: W1203 12:59:47.275491 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28e77ac6_94cf_4a90_b75f_f26c786cb2c6.slice/crio-e817dc4c9519b5dec857a205ed1242659e176dddfc3dd5b8fa04c36af8e2b3d4 WatchSource:0}: Error finding container e817dc4c9519b5dec857a205ed1242659e176dddfc3dd5b8fa04c36af8e2b3d4: Status 404 returned error can't find the container with id e817dc4c9519b5dec857a205ed1242659e176dddfc3dd5b8fa04c36af8e2b3d4 Dec 03 12:59:47 crc kubenswrapper[4990]: I1203 12:59:47.482963 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.104225 4990 generic.go:334] "Generic (PLEG): container finished" podID="1c5047b8-3306-40b5-bbaf-a30c6286662b" containerID="1cb1b6ac4dba6a6576c3d28e7424f916b234bbcd8e4ec0c5bedbab7ddc617e8e" exitCode=0 Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.104310 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rx7m6" event={"ID":"1c5047b8-3306-40b5-bbaf-a30c6286662b","Type":"ContainerDied","Data":"1cb1b6ac4dba6a6576c3d28e7424f916b234bbcd8e4ec0c5bedbab7ddc617e8e"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.108437 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"e94695d94cd07336f965574eb9e5b5d9db6bf6d265e5389bb7b80aff339030ce"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.113734 4990 generic.go:334] "Generic (PLEG): container finished" podID="95aafce7-81a5-45a4-aa39-82431c05b665" containerID="7648fd6cf60c1497abcd8aeafd954a237266279d769c2363692d2001a1b4ab39" exitCode=0 Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.113792 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kkvhj" event={"ID":"95aafce7-81a5-45a4-aa39-82431c05b665","Type":"ContainerDied","Data":"7648fd6cf60c1497abcd8aeafd954a237266279d769c2363692d2001a1b4ab39"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.115860 4990 generic.go:334] "Generic (PLEG): container finished" podID="2e3879fb-c88c-4c57-9620-d620812aa5cf" containerID="24ca0561482fd601a76f310fd5fdaad86b52740bd4c9f6799180d6d2bd09dc0b" exitCode=0 Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.115914 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5066-account-create-update-w7bck" event={"ID":"2e3879fb-c88c-4c57-9620-d620812aa5cf","Type":"ContainerDied","Data":"24ca0561482fd601a76f310fd5fdaad86b52740bd4c9f6799180d6d2bd09dc0b"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.118264 4990 generic.go:334] "Generic (PLEG): container finished" podID="20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" containerID="03aefbc52bf21d9e184d73edc83eaa15b8a3e528bf29d6af92fa0dafbf8b49ca" exitCode=0 Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.118506 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d8c5-account-create-update-qhz9j" event={"ID":"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7","Type":"ContainerDied","Data":"03aefbc52bf21d9e184d73edc83eaa15b8a3e528bf29d6af92fa0dafbf8b49ca"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.120193 4990 generic.go:334] "Generic (PLEG): container finished" podID="28e77ac6-94cf-4a90-b75f-f26c786cb2c6" containerID="80a36ee5d21d345ea1e61546d3b9de599f81caccc7d2905415897e1f22c630ab" exitCode=0 Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.120253 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-29e5-account-create-update-t945b" event={"ID":"28e77ac6-94cf-4a90-b75f-f26c786cb2c6","Type":"ContainerDied","Data":"80a36ee5d21d345ea1e61546d3b9de599f81caccc7d2905415897e1f22c630ab"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.120275 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-29e5-account-create-update-t945b" event={"ID":"28e77ac6-94cf-4a90-b75f-f26c786cb2c6","Type":"ContainerStarted","Data":"e817dc4c9519b5dec857a205ed1242659e176dddfc3dd5b8fa04c36af8e2b3d4"} Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.124135 4990 generic.go:334] "Generic (PLEG): container finished" podID="45d0ab64-a5d0-4e19-8b39-d38906823542" containerID="6d485fea3634ef62dde1cc4122194d9c199265ff2dc7e0fa5e0275139fa116c3" exitCode=0 Dec 03 12:59:48 crc kubenswrapper[4990]: I1203 12:59:48.124197 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjl5k" event={"ID":"45d0ab64-a5d0-4e19-8b39-d38906823542","Type":"ContainerDied","Data":"6d485fea3634ef62dde1cc4122194d9c199265ff2dc7e0fa5e0275139fa116c3"} Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.145986 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"3d466739b904422e857bb4080544bced0a2ada69f2b281cf7aae5a976bf788b3"} Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.357695 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-pfvtg"] Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.358975 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.362675 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.362849 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t28jn" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.363434 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.363519 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.374047 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-flkbm"] Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.375729 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.378283 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.379757 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gzpbp" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.388657 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-flkbm"] Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.400827 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pfvtg"] Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.409949 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2m96\" (UniqueName: \"kubernetes.io/projected/8f5bbad8-f863-461f-a580-3989b7d2ceaa-kube-api-access-x2m96\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.410001 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-config-data\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.410037 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-config-data\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.410352 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-db-sync-config-data\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.410564 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-combined-ca-bundle\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.410615 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-combined-ca-bundle\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.410729 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvd5f\" (UniqueName: \"kubernetes.io/projected/fb56fae9-35bb-4e75-95b3-6ba3185189aa-kube-api-access-tvd5f\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514016 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-db-sync-config-data\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514472 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-combined-ca-bundle\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514501 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-combined-ca-bundle\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514577 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvd5f\" (UniqueName: \"kubernetes.io/projected/fb56fae9-35bb-4e75-95b3-6ba3185189aa-kube-api-access-tvd5f\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514765 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2m96\" (UniqueName: \"kubernetes.io/projected/8f5bbad8-f863-461f-a580-3989b7d2ceaa-kube-api-access-x2m96\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514828 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-config-data\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.514911 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-config-data\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.522849 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-config-data\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.526965 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-config-data\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.534243 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-combined-ca-bundle\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.534788 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvd5f\" (UniqueName: \"kubernetes.io/projected/fb56fae9-35bb-4e75-95b3-6ba3185189aa-kube-api-access-tvd5f\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.549091 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-db-sync-config-data\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.554832 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2m96\" (UniqueName: \"kubernetes.io/projected/8f5bbad8-f863-461f-a580-3989b7d2ceaa-kube-api-access-x2m96\") pod \"glance-db-sync-flkbm\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.580607 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-combined-ca-bundle\") pod \"keystone-db-sync-pfvtg\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.658248 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.686531 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pfvtg" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.714006 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-flkbm" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.726866 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsl8d\" (UniqueName: \"kubernetes.io/projected/1c5047b8-3306-40b5-bbaf-a30c6286662b-kube-api-access-fsl8d\") pod \"1c5047b8-3306-40b5-bbaf-a30c6286662b\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.727209 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c5047b8-3306-40b5-bbaf-a30c6286662b-operator-scripts\") pod \"1c5047b8-3306-40b5-bbaf-a30c6286662b\" (UID: \"1c5047b8-3306-40b5-bbaf-a30c6286662b\") " Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.730415 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c5047b8-3306-40b5-bbaf-a30c6286662b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c5047b8-3306-40b5-bbaf-a30c6286662b" (UID: "1c5047b8-3306-40b5-bbaf-a30c6286662b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.749134 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c5047b8-3306-40b5-bbaf-a30c6286662b-kube-api-access-fsl8d" (OuterVolumeSpecName: "kube-api-access-fsl8d") pod "1c5047b8-3306-40b5-bbaf-a30c6286662b" (UID: "1c5047b8-3306-40b5-bbaf-a30c6286662b"). InnerVolumeSpecName "kube-api-access-fsl8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.758347 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c5047b8-3306-40b5-bbaf-a30c6286662b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.758410 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsl8d\" (UniqueName: \"kubernetes.io/projected/1c5047b8-3306-40b5-bbaf-a30c6286662b-kube-api-access-fsl8d\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.856955 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.962759 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhxg2\" (UniqueName: \"kubernetes.io/projected/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-kube-api-access-nhxg2\") pod \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.962900 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-operator-scripts\") pod \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\" (UID: \"28e77ac6-94cf-4a90-b75f-f26c786cb2c6\") " Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.964606 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "28e77ac6-94cf-4a90-b75f-f26c786cb2c6" (UID: "28e77ac6-94cf-4a90-b75f-f26c786cb2c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.965055 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:49 crc kubenswrapper[4990]: I1203 12:59:49.972601 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-kube-api-access-nhxg2" (OuterVolumeSpecName: "kube-api-access-nhxg2") pod "28e77ac6-94cf-4a90-b75f-f26c786cb2c6" (UID: "28e77ac6-94cf-4a90-b75f-f26c786cb2c6"). InnerVolumeSpecName "kube-api-access-nhxg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.040862 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.066275 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhxg2\" (UniqueName: \"kubernetes.io/projected/28e77ac6-94cf-4a90-b75f-f26c786cb2c6-kube-api-access-nhxg2\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.066857 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.084224 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.091430 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169032 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95aafce7-81a5-45a4-aa39-82431c05b665-operator-scripts\") pod \"95aafce7-81a5-45a4-aa39-82431c05b665\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169218 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z4kr\" (UniqueName: \"kubernetes.io/projected/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-kube-api-access-8z4kr\") pod \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169307 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-operator-scripts\") pod \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\" (UID: \"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169370 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp4rl\" (UniqueName: \"kubernetes.io/projected/2e3879fb-c88c-4c57-9620-d620812aa5cf-kube-api-access-rp4rl\") pod \"2e3879fb-c88c-4c57-9620-d620812aa5cf\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169395 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3879fb-c88c-4c57-9620-d620812aa5cf-operator-scripts\") pod \"2e3879fb-c88c-4c57-9620-d620812aa5cf\" (UID: \"2e3879fb-c88c-4c57-9620-d620812aa5cf\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169517 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txn2l\" (UniqueName: \"kubernetes.io/projected/45d0ab64-a5d0-4e19-8b39-d38906823542-kube-api-access-txn2l\") pod \"45d0ab64-a5d0-4e19-8b39-d38906823542\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169582 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d0ab64-a5d0-4e19-8b39-d38906823542-operator-scripts\") pod \"45d0ab64-a5d0-4e19-8b39-d38906823542\" (UID: \"45d0ab64-a5d0-4e19-8b39-d38906823542\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.169628 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn857\" (UniqueName: \"kubernetes.io/projected/95aafce7-81a5-45a4-aa39-82431c05b665-kube-api-access-pn857\") pod \"95aafce7-81a5-45a4-aa39-82431c05b665\" (UID: \"95aafce7-81a5-45a4-aa39-82431c05b665\") " Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.171221 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95aafce7-81a5-45a4-aa39-82431c05b665-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95aafce7-81a5-45a4-aa39-82431c05b665" (UID: "95aafce7-81a5-45a4-aa39-82431c05b665"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.173988 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d0ab64-a5d0-4e19-8b39-d38906823542-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45d0ab64-a5d0-4e19-8b39-d38906823542" (UID: "45d0ab64-a5d0-4e19-8b39-d38906823542"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.174756 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" (UID: "20984793-ac7c-4ef4-bbf2-cb6e21d03bc7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.174923 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e3879fb-c88c-4c57-9620-d620812aa5cf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e3879fb-c88c-4c57-9620-d620812aa5cf" (UID: "2e3879fb-c88c-4c57-9620-d620812aa5cf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.175141 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-29e5-account-create-update-t945b" event={"ID":"28e77ac6-94cf-4a90-b75f-f26c786cb2c6","Type":"ContainerDied","Data":"e817dc4c9519b5dec857a205ed1242659e176dddfc3dd5b8fa04c36af8e2b3d4"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.175175 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e817dc4c9519b5dec857a205ed1242659e176dddfc3dd5b8fa04c36af8e2b3d4" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.175247 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29e5-account-create-update-t945b" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.176168 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-kube-api-access-8z4kr" (OuterVolumeSpecName: "kube-api-access-8z4kr") pod "20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" (UID: "20984793-ac7c-4ef4-bbf2-cb6e21d03bc7"). InnerVolumeSpecName "kube-api-access-8z4kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.183786 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjl5k" event={"ID":"45d0ab64-a5d0-4e19-8b39-d38906823542","Type":"ContainerDied","Data":"1c9a51f8ab27d1559bc31c0effe779cebdfb724dfe08ea8fe6f9bafc0bcb84f3"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.183843 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c9a51f8ab27d1559bc31c0effe779cebdfb724dfe08ea8fe6f9bafc0bcb84f3" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.184007 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjl5k" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.186772 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rx7m6" event={"ID":"1c5047b8-3306-40b5-bbaf-a30c6286662b","Type":"ContainerDied","Data":"12fd6adad3a25ced5d0c4c136532e4c306b56625e50e8bae2da34d4627866adb"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.186800 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12fd6adad3a25ced5d0c4c136532e4c306b56625e50e8bae2da34d4627866adb" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.186872 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rx7m6" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.228624 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e3879fb-c88c-4c57-9620-d620812aa5cf-kube-api-access-rp4rl" (OuterVolumeSpecName: "kube-api-access-rp4rl") pod "2e3879fb-c88c-4c57-9620-d620812aa5cf" (UID: "2e3879fb-c88c-4c57-9620-d620812aa5cf"). InnerVolumeSpecName "kube-api-access-rp4rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.239277 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95aafce7-81a5-45a4-aa39-82431c05b665-kube-api-access-pn857" (OuterVolumeSpecName: "kube-api-access-pn857") pod "95aafce7-81a5-45a4-aa39-82431c05b665" (UID: "95aafce7-81a5-45a4-aa39-82431c05b665"). InnerVolumeSpecName "kube-api-access-pn857". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.239638 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d0ab64-a5d0-4e19-8b39-d38906823542-kube-api-access-txn2l" (OuterVolumeSpecName: "kube-api-access-txn2l") pod "45d0ab64-a5d0-4e19-8b39-d38906823542" (UID: "45d0ab64-a5d0-4e19-8b39-d38906823542"). InnerVolumeSpecName "kube-api-access-txn2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.261370 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"85efea0b14c4eafd66fbb2815a2eec9c370150233ad848bac32fefd2ebb6f8ad"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.261422 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"0c6e9d01f1f85a3ee72e9a348b04ef5b661e8ea7335657eba0b781b35ec9236a"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272119 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn857\" (UniqueName: \"kubernetes.io/projected/95aafce7-81a5-45a4-aa39-82431c05b665-kube-api-access-pn857\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272152 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95aafce7-81a5-45a4-aa39-82431c05b665-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272165 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z4kr\" (UniqueName: \"kubernetes.io/projected/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-kube-api-access-8z4kr\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272175 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272186 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp4rl\" (UniqueName: \"kubernetes.io/projected/2e3879fb-c88c-4c57-9620-d620812aa5cf-kube-api-access-rp4rl\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272196 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e3879fb-c88c-4c57-9620-d620812aa5cf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272206 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txn2l\" (UniqueName: \"kubernetes.io/projected/45d0ab64-a5d0-4e19-8b39-d38906823542-kube-api-access-txn2l\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.272217 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d0ab64-a5d0-4e19-8b39-d38906823542-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.278885 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kkvhj" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.285145 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5066-account-create-update-w7bck" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.314685 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d8c5-account-create-update-qhz9j" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.317921 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kkvhj" event={"ID":"95aafce7-81a5-45a4-aa39-82431c05b665","Type":"ContainerDied","Data":"7719b4a1826e3ef3bdc324a3561354073c6fa7ada94eaf7caf343510a751b1e5"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.317961 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7719b4a1826e3ef3bdc324a3561354073c6fa7ada94eaf7caf343510a751b1e5" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.317977 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5066-account-create-update-w7bck" event={"ID":"2e3879fb-c88c-4c57-9620-d620812aa5cf","Type":"ContainerDied","Data":"4aff73f8877276a0f8632941b829e74bb349868dbfdb98e8354b539d210d2045"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.317988 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4aff73f8877276a0f8632941b829e74bb349868dbfdb98e8354b539d210d2045" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.317998 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d8c5-account-create-update-qhz9j" event={"ID":"20984793-ac7c-4ef4-bbf2-cb6e21d03bc7","Type":"ContainerDied","Data":"4285c328bd0d853460ef5ada0b3e625a920d091d4aa543f4a6ec6578ea364c68"} Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.318008 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4285c328bd0d853460ef5ada0b3e625a920d091d4aa543f4a6ec6578ea364c68" Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.364183 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pfvtg"] Dec 03 12:59:50 crc kubenswrapper[4990]: I1203 12:59:50.523054 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-flkbm"] Dec 03 12:59:50 crc kubenswrapper[4990]: W1203 12:59:50.525839 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f5bbad8_f863_461f_a580_3989b7d2ceaa.slice/crio-e5ed9166d690f3232548c8bd8dbfb87f80b0d54e08bac856da70dda754882815 WatchSource:0}: Error finding container e5ed9166d690f3232548c8bd8dbfb87f80b0d54e08bac856da70dda754882815: Status 404 returned error can't find the container with id e5ed9166d690f3232548c8bd8dbfb87f80b0d54e08bac856da70dda754882815 Dec 03 12:59:51 crc kubenswrapper[4990]: I1203 12:59:51.328782 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pfvtg" event={"ID":"fb56fae9-35bb-4e75-95b3-6ba3185189aa","Type":"ContainerStarted","Data":"e91dd276d08d77e8b54840809c0e991bb6c5f8094934c96207670106b50f35a8"} Dec 03 12:59:51 crc kubenswrapper[4990]: I1203 12:59:51.331408 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-flkbm" event={"ID":"8f5bbad8-f863-461f-a580-3989b7d2ceaa","Type":"ContainerStarted","Data":"e5ed9166d690f3232548c8bd8dbfb87f80b0d54e08bac856da70dda754882815"} Dec 03 12:59:51 crc kubenswrapper[4990]: I1203 12:59:51.334242 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"5d9f8bcd7e84cb1e0e7d893b8998337a377d8664e2c2578001aa0ba8797dd5cf"} Dec 03 12:59:52 crc kubenswrapper[4990]: I1203 12:59:52.382652 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"07cf11fc530c554a9cfbf33e40bbebd6dea0c446f38071af078ee92006063052"} Dec 03 12:59:53 crc kubenswrapper[4990]: I1203 12:59:53.409917 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"91f87b86eeb6df78e2a5f2917bdcde69d7b6abd0198943222b3ec4fe275f73ae"} Dec 03 12:59:55 crc kubenswrapper[4990]: I1203 12:59:55.441731 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"70873077d148fd4556bd1614e19bedc4c17a1bd10196e5bb5cca3bdd61ca35fb"} Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.133782 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc"] Dec 03 13:00:00 crc kubenswrapper[4990]: E1203 13:00:00.134425 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d0ab64-a5d0-4e19-8b39-d38906823542" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134437 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d0ab64-a5d0-4e19-8b39-d38906823542" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: E1203 13:00:00.134494 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134504 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: E1203 13:00:00.134516 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3879fb-c88c-4c57-9620-d620812aa5cf" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134525 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3879fb-c88c-4c57-9620-d620812aa5cf" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: E1203 13:00:00.134535 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e77ac6-94cf-4a90-b75f-f26c786cb2c6" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134542 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e77ac6-94cf-4a90-b75f-f26c786cb2c6" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: E1203 13:00:00.134558 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95aafce7-81a5-45a4-aa39-82431c05b665" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134564 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="95aafce7-81a5-45a4-aa39-82431c05b665" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: E1203 13:00:00.134575 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5047b8-3306-40b5-bbaf-a30c6286662b" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134583 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5047b8-3306-40b5-bbaf-a30c6286662b" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134755 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="95aafce7-81a5-45a4-aa39-82431c05b665" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134784 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d0ab64-a5d0-4e19-8b39-d38906823542" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134798 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e77ac6-94cf-4a90-b75f-f26c786cb2c6" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134808 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5047b8-3306-40b5-bbaf-a30c6286662b" containerName="mariadb-database-create" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134820 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.134830 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3879fb-c88c-4c57-9620-d620812aa5cf" containerName="mariadb-account-create-update" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.135385 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.138259 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.138509 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.145222 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc"] Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.248475 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4q24\" (UniqueName: \"kubernetes.io/projected/7f73e432-c318-45bf-a8ec-93b26f1f823d-kube-api-access-f4q24\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.248862 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f73e432-c318-45bf-a8ec-93b26f1f823d-config-volume\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.248906 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f73e432-c318-45bf-a8ec-93b26f1f823d-secret-volume\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.350420 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4q24\" (UniqueName: \"kubernetes.io/projected/7f73e432-c318-45bf-a8ec-93b26f1f823d-kube-api-access-f4q24\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.350569 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f73e432-c318-45bf-a8ec-93b26f1f823d-config-volume\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.350613 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f73e432-c318-45bf-a8ec-93b26f1f823d-secret-volume\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.351894 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f73e432-c318-45bf-a8ec-93b26f1f823d-config-volume\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.363559 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f73e432-c318-45bf-a8ec-93b26f1f823d-secret-volume\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.369241 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4q24\" (UniqueName: \"kubernetes.io/projected/7f73e432-c318-45bf-a8ec-93b26f1f823d-kube-api-access-f4q24\") pod \"collect-profiles-29412780-vrpqc\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:00 crc kubenswrapper[4990]: I1203 13:00:00.454148 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:03 crc kubenswrapper[4990]: I1203 13:00:03.286772 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:00:03 crc kubenswrapper[4990]: I1203 13:00:03.287552 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:00:03 crc kubenswrapper[4990]: I1203 13:00:03.287608 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:00:03 crc kubenswrapper[4990]: I1203 13:00:03.288505 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af393e5087a271fce924a6dbdf31ad1b2b0bea5f1a2211228838f21a6a0541a1"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:00:03 crc kubenswrapper[4990]: I1203 13:00:03.288566 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://af393e5087a271fce924a6dbdf31ad1b2b0bea5f1a2211228838f21a6a0541a1" gracePeriod=600 Dec 03 13:00:04 crc kubenswrapper[4990]: I1203 13:00:04.781774 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-dtt26" podUID="20a68e4a-3638-430c-a27b-7d535613c3d3" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 03 13:00:15 crc kubenswrapper[4990]: I1203 13:00:15.616104 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="af393e5087a271fce924a6dbdf31ad1b2b0bea5f1a2211228838f21a6a0541a1" exitCode=0 Dec 03 13:00:15 crc kubenswrapper[4990]: I1203 13:00:15.616175 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"af393e5087a271fce924a6dbdf31ad1b2b0bea5f1a2211228838f21a6a0541a1"} Dec 03 13:00:15 crc kubenswrapper[4990]: I1203 13:00:15.616749 4990 scope.go:117] "RemoveContainer" containerID="80c0112fe475b4f4e2b7d1d072f2e686f621ee22a4d45518867b967855b8112b" Dec 03 13:00:16 crc kubenswrapper[4990]: E1203 13:00:16.808656 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 03 13:00:16 crc kubenswrapper[4990]: E1203 13:00:16.809167 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x2m96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-flkbm_openstack(8f5bbad8-f863-461f-a580-3989b7d2ceaa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:00:16 crc kubenswrapper[4990]: E1203 13:00:16.810370 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-flkbm" podUID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" Dec 03 13:00:17 crc kubenswrapper[4990]: E1203 13:00:17.634108 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-flkbm" podUID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" Dec 03 13:00:17 crc kubenswrapper[4990]: W1203 13:00:17.952931 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f73e432_c318_45bf_a8ec_93b26f1f823d.slice/crio-fa6ee227c051732940d040b3dcb51114c3a20e0b85cbad977772958ffd239a93 WatchSource:0}: Error finding container fa6ee227c051732940d040b3dcb51114c3a20e0b85cbad977772958ffd239a93: Status 404 returned error can't find the container with id fa6ee227c051732940d040b3dcb51114c3a20e0b85cbad977772958ffd239a93 Dec 03 13:00:17 crc kubenswrapper[4990]: I1203 13:00:17.956778 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc"] Dec 03 13:00:18 crc kubenswrapper[4990]: I1203 13:00:18.654901 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"946672e66bea617e7661c0dc20d08f592bebd599de9cfe4c9155f58c366ccbca"} Dec 03 13:00:18 crc kubenswrapper[4990]: I1203 13:00:18.658155 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pfvtg" event={"ID":"fb56fae9-35bb-4e75-95b3-6ba3185189aa","Type":"ContainerStarted","Data":"b8b001866ba17d9f295eb7f115c2e6fe780b3421d5838e9b6e5d141dc09eb30b"} Dec 03 13:00:18 crc kubenswrapper[4990]: I1203 13:00:18.665320 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" event={"ID":"7f73e432-c318-45bf-a8ec-93b26f1f823d","Type":"ContainerStarted","Data":"5f8ce297f026d7e58d417b5143bfebe67c674959542e044a98564069411d56f0"} Dec 03 13:00:18 crc kubenswrapper[4990]: I1203 13:00:18.665361 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" event={"ID":"7f73e432-c318-45bf-a8ec-93b26f1f823d","Type":"ContainerStarted","Data":"fa6ee227c051732940d040b3dcb51114c3a20e0b85cbad977772958ffd239a93"} Dec 03 13:00:18 crc kubenswrapper[4990]: I1203 13:00:18.698986 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-pfvtg" podStartSLOduration=2.7290268060000002 podStartE2EDuration="29.698964543s" podCreationTimestamp="2025-12-03 12:59:49 +0000 UTC" firstStartedPulling="2025-12-03 12:59:50.488760781 +0000 UTC m=+1338.630672030" lastFinishedPulling="2025-12-03 13:00:17.458698538 +0000 UTC m=+1365.600609767" observedRunningTime="2025-12-03 13:00:18.691859216 +0000 UTC m=+1366.833770455" watchObservedRunningTime="2025-12-03 13:00:18.698964543 +0000 UTC m=+1366.840875772" Dec 03 13:00:19 crc kubenswrapper[4990]: I1203 13:00:19.677557 4990 generic.go:334] "Generic (PLEG): container finished" podID="7f73e432-c318-45bf-a8ec-93b26f1f823d" containerID="5f8ce297f026d7e58d417b5143bfebe67c674959542e044a98564069411d56f0" exitCode=0 Dec 03 13:00:19 crc kubenswrapper[4990]: I1203 13:00:19.677830 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" event={"ID":"7f73e432-c318-45bf-a8ec-93b26f1f823d","Type":"ContainerDied","Data":"5f8ce297f026d7e58d417b5143bfebe67c674959542e044a98564069411d56f0"} Dec 03 13:00:19 crc kubenswrapper[4990]: I1203 13:00:19.681203 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.081002 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.255104 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f73e432-c318-45bf-a8ec-93b26f1f823d-secret-volume\") pod \"7f73e432-c318-45bf-a8ec-93b26f1f823d\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.255186 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4q24\" (UniqueName: \"kubernetes.io/projected/7f73e432-c318-45bf-a8ec-93b26f1f823d-kube-api-access-f4q24\") pod \"7f73e432-c318-45bf-a8ec-93b26f1f823d\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.255323 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f73e432-c318-45bf-a8ec-93b26f1f823d-config-volume\") pod \"7f73e432-c318-45bf-a8ec-93b26f1f823d\" (UID: \"7f73e432-c318-45bf-a8ec-93b26f1f823d\") " Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.256187 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f73e432-c318-45bf-a8ec-93b26f1f823d-config-volume" (OuterVolumeSpecName: "config-volume") pod "7f73e432-c318-45bf-a8ec-93b26f1f823d" (UID: "7f73e432-c318-45bf-a8ec-93b26f1f823d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.262578 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f73e432-c318-45bf-a8ec-93b26f1f823d-kube-api-access-f4q24" (OuterVolumeSpecName: "kube-api-access-f4q24") pod "7f73e432-c318-45bf-a8ec-93b26f1f823d" (UID: "7f73e432-c318-45bf-a8ec-93b26f1f823d"). InnerVolumeSpecName "kube-api-access-f4q24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.263011 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f73e432-c318-45bf-a8ec-93b26f1f823d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7f73e432-c318-45bf-a8ec-93b26f1f823d" (UID: "7f73e432-c318-45bf-a8ec-93b26f1f823d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.358072 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f73e432-c318-45bf-a8ec-93b26f1f823d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.358129 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f73e432-c318-45bf-a8ec-93b26f1f823d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.358140 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4q24\" (UniqueName: \"kubernetes.io/projected/7f73e432-c318-45bf-a8ec-93b26f1f823d-kube-api-access-f4q24\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.718174 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" event={"ID":"7f73e432-c318-45bf-a8ec-93b26f1f823d","Type":"ContainerDied","Data":"fa6ee227c051732940d040b3dcb51114c3a20e0b85cbad977772958ffd239a93"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.718431 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa6ee227c051732940d040b3dcb51114c3a20e0b85cbad977772958ffd239a93" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.718272 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc" Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.727656 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"1653d9ad256d8fd4d12694ec20d946a552054711d8904c2c619e1d013fc01e1d"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.727709 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"a180f811b91fd0bb8229875fcd2a4bbb027911155ca9acb98f566425701db3d0"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.727726 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"f5bf4034ee7a8e126b0459979259a97c3ffd1a7842d77c80ee8bc8b39b089fb6"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.727737 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"c6f168aa91a4567be48e2b24e57d858bdbfab0ca87932b2c232c3ccc12783698"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.727750 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"c77b3c6f3cd1285b3a065619bf527a653a77ad6553a8599422b31b68515802ce"} Dec 03 13:00:21 crc kubenswrapper[4990]: I1203 13:00:21.727763 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"bcda88079fdd9fcc423de223bb1ef68d70627f3d9f8dc4612a006471366a11f9"} Dec 03 13:00:22 crc kubenswrapper[4990]: I1203 13:00:22.741757 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f7d434c5-0459-4d5a-a401-4a0c4b82a553","Type":"ContainerStarted","Data":"86ea58018f025005cc09d74313ba835ed989a10b5aee8c6e139cb3b7d34e4c84"} Dec 03 13:00:22 crc kubenswrapper[4990]: I1203 13:00:22.783595 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=68.793637967 podStartE2EDuration="1m41.783568466s" podCreationTimestamp="2025-12-03 12:58:41 +0000 UTC" firstStartedPulling="2025-12-03 12:59:47.50216756 +0000 UTC m=+1335.644078789" lastFinishedPulling="2025-12-03 13:00:20.492098059 +0000 UTC m=+1368.634009288" observedRunningTime="2025-12-03 13:00:22.774731352 +0000 UTC m=+1370.916642581" watchObservedRunningTime="2025-12-03 13:00:22.783568466 +0000 UTC m=+1370.925479705" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.102837 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-bntzq"] Dec 03 13:00:23 crc kubenswrapper[4990]: E1203 13:00:23.103396 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f73e432-c318-45bf-a8ec-93b26f1f823d" containerName="collect-profiles" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.103413 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f73e432-c318-45bf-a8ec-93b26f1f823d" containerName="collect-profiles" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.103599 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f73e432-c318-45bf-a8ec-93b26f1f823d" containerName="collect-profiles" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.104580 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.111892 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.122689 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-bntzq"] Dec 03 13:00:23 crc kubenswrapper[4990]: E1203 13:00:23.187145 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb56fae9_35bb_4e75_95b3_6ba3185189aa.slice/crio-b8b001866ba17d9f295eb7f115c2e6fe780b3421d5838e9b6e5d141dc09eb30b.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.195985 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.196080 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.196146 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.196180 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lz2m\" (UniqueName: \"kubernetes.io/projected/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-kube-api-access-9lz2m\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.196199 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.196237 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-config\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.298121 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.298222 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.298268 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lz2m\" (UniqueName: \"kubernetes.io/projected/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-kube-api-access-9lz2m\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.298330 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.298390 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-config\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.298432 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.299147 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.299253 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.299254 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.299656 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.300929 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-config\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.320864 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lz2m\" (UniqueName: \"kubernetes.io/projected/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-kube-api-access-9lz2m\") pod \"dnsmasq-dns-6d5b6d6b67-bntzq\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.426553 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.750488 4990 generic.go:334] "Generic (PLEG): container finished" podID="fb56fae9-35bb-4e75-95b3-6ba3185189aa" containerID="b8b001866ba17d9f295eb7f115c2e6fe780b3421d5838e9b6e5d141dc09eb30b" exitCode=0 Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.750557 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pfvtg" event={"ID":"fb56fae9-35bb-4e75-95b3-6ba3185189aa","Type":"ContainerDied","Data":"b8b001866ba17d9f295eb7f115c2e6fe780b3421d5838e9b6e5d141dc09eb30b"} Dec 03 13:00:23 crc kubenswrapper[4990]: I1203 13:00:23.875393 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-bntzq"] Dec 03 13:00:24 crc kubenswrapper[4990]: I1203 13:00:24.760093 4990 generic.go:334] "Generic (PLEG): container finished" podID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerID="92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390" exitCode=0 Dec 03 13:00:24 crc kubenswrapper[4990]: I1203 13:00:24.760169 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" event={"ID":"bfb6ffdc-2484-46fe-ad9d-69a60125d20f","Type":"ContainerDied","Data":"92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390"} Dec 03 13:00:24 crc kubenswrapper[4990]: I1203 13:00:24.760483 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" event={"ID":"bfb6ffdc-2484-46fe-ad9d-69a60125d20f","Type":"ContainerStarted","Data":"36791013bce3e0f80366d7486304719a4b592ea700ca369a5ba07d42fec6f313"} Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.228934 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pfvtg" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.414299 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvd5f\" (UniqueName: \"kubernetes.io/projected/fb56fae9-35bb-4e75-95b3-6ba3185189aa-kube-api-access-tvd5f\") pod \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.414674 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-config-data\") pod \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.414758 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-combined-ca-bundle\") pod \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\" (UID: \"fb56fae9-35bb-4e75-95b3-6ba3185189aa\") " Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.419883 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb56fae9-35bb-4e75-95b3-6ba3185189aa-kube-api-access-tvd5f" (OuterVolumeSpecName: "kube-api-access-tvd5f") pod "fb56fae9-35bb-4e75-95b3-6ba3185189aa" (UID: "fb56fae9-35bb-4e75-95b3-6ba3185189aa"). InnerVolumeSpecName "kube-api-access-tvd5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.440969 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb56fae9-35bb-4e75-95b3-6ba3185189aa" (UID: "fb56fae9-35bb-4e75-95b3-6ba3185189aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.461360 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-config-data" (OuterVolumeSpecName: "config-data") pod "fb56fae9-35bb-4e75-95b3-6ba3185189aa" (UID: "fb56fae9-35bb-4e75-95b3-6ba3185189aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.516820 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.516860 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvd5f\" (UniqueName: \"kubernetes.io/projected/fb56fae9-35bb-4e75-95b3-6ba3185189aa-kube-api-access-tvd5f\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.516872 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb56fae9-35bb-4e75-95b3-6ba3185189aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.782927 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" event={"ID":"bfb6ffdc-2484-46fe-ad9d-69a60125d20f","Type":"ContainerStarted","Data":"eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff"} Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.784250 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.787586 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pfvtg" event={"ID":"fb56fae9-35bb-4e75-95b3-6ba3185189aa","Type":"ContainerDied","Data":"e91dd276d08d77e8b54840809c0e991bb6c5f8094934c96207670106b50f35a8"} Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.787636 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pfvtg" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.787649 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e91dd276d08d77e8b54840809c0e991bb6c5f8094934c96207670106b50f35a8" Dec 03 13:00:25 crc kubenswrapper[4990]: I1203 13:00:25.827339 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" podStartSLOduration=2.82731786 podStartE2EDuration="2.82731786s" podCreationTimestamp="2025-12-03 13:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:00:25.815858927 +0000 UTC m=+1373.957770176" watchObservedRunningTime="2025-12-03 13:00:25.82731786 +0000 UTC m=+1373.969229099" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.058207 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-bntzq"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.113027 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mmj4c"] Dec 03 13:00:26 crc kubenswrapper[4990]: E1203 13:00:26.117894 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb56fae9-35bb-4e75-95b3-6ba3185189aa" containerName="keystone-db-sync" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.118164 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb56fae9-35bb-4e75-95b3-6ba3185189aa" containerName="keystone-db-sync" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.118572 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb56fae9-35bb-4e75-95b3-6ba3185189aa" containerName="keystone-db-sync" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.119301 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.123992 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.124201 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t28jn" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.124016 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.124556 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.136805 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.148357 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-vc5ws"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.150087 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.182346 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mmj4c"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.219277 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-vc5ws"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.231176 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z56d9\" (UniqueName: \"kubernetes.io/projected/fc091b50-7b0c-4922-bbf1-d203a982277b-kube-api-access-z56d9\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.235975 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-fernet-keys\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.236030 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-combined-ca-bundle\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.236064 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-credential-keys\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.236397 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-scripts\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.236585 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-config-data\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366177 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-scripts\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366252 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366293 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-config-data\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366318 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps2fm\" (UniqueName: \"kubernetes.io/projected/13635c51-8af0-43cf-a0bc-bce71a531b47-kube-api-access-ps2fm\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366369 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-config\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366426 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366738 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z56d9\" (UniqueName: \"kubernetes.io/projected/fc091b50-7b0c-4922-bbf1-d203a982277b-kube-api-access-z56d9\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366808 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366853 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-fernet-keys\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366878 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366909 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-combined-ca-bundle\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.366934 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-credential-keys\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.379119 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-scripts\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.381294 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-combined-ca-bundle\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.383185 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-credential-keys\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.386258 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-fernet-keys\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.410574 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-config-data\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.428350 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z56d9\" (UniqueName: \"kubernetes.io/projected/fc091b50-7b0c-4922-bbf1-d203a982277b-kube-api-access-z56d9\") pod \"keystone-bootstrap-mmj4c\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.477570 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.477639 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps2fm\" (UniqueName: \"kubernetes.io/projected/13635c51-8af0-43cf-a0bc-bce71a531b47-kube-api-access-ps2fm\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.477681 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-config\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.477718 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.477756 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.477804 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.478910 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.479050 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.484831 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-psk4l"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.486281 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.486387 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.499946 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.503599 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.504309 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wsxks" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.507259 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.509529 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-config\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.509636 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-psk4l"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.510142 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.531239 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c484f75c7-xh677"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.533121 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.545274 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps2fm\" (UniqueName: \"kubernetes.io/projected/13635c51-8af0-43cf-a0bc-bce71a531b47-kube-api-access-ps2fm\") pod \"dnsmasq-dns-6f8c45789f-vc5ws\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.545790 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-6hgcg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.546148 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.546302 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.550104 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.554871 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-cqrdj"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.558248 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.568226 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nxxxj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.568587 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.568741 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.575136 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c484f75c7-xh677"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.613532 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cqrdj"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.651508 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2kjtg"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.652773 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.667835 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.668217 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-h9hc9" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.680965 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-db-sync-config-data\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681473 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlgh9\" (UniqueName: \"kubernetes.io/projected/81f50f60-4235-4365-84dd-3cfcf9142ae4-kube-api-access-tlgh9\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681518 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-combined-ca-bundle\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681545 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-scripts\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681566 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-config-data\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681589 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-logs\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681605 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-config\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681654 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkwqs\" (UniqueName: \"kubernetes.io/projected/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-kube-api-access-bkwqs\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681673 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-scripts\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681707 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81f50f60-4235-4365-84dd-3cfcf9142ae4-etc-machine-id\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681729 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-combined-ca-bundle\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681746 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-config-data\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681761 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlt7q\" (UniqueName: \"kubernetes.io/projected/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-kube-api-access-xlt7q\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681811 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-combined-ca-bundle\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681839 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-db-sync-config-data\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681860 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88czv\" (UniqueName: \"kubernetes.io/projected/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-kube-api-access-88czv\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.681900 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-horizon-secret-key\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.689366 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2kjtg"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.723525 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-9n7xj"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.724991 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.733933 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.734117 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.734662 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n66n7" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.749093 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-vc5ws"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.749983 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.784889 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkwqs\" (UniqueName: \"kubernetes.io/projected/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-kube-api-access-bkwqs\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.784944 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-scripts\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.784991 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-config-data\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785023 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-combined-ca-bundle\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785050 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-scripts\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785105 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81f50f60-4235-4365-84dd-3cfcf9142ae4-etc-machine-id\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785129 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-combined-ca-bundle\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785360 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81f50f60-4235-4365-84dd-3cfcf9142ae4-etc-machine-id\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785157 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-config-data\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785713 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlt7q\" (UniqueName: \"kubernetes.io/projected/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-kube-api-access-xlt7q\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785782 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsnx6\" (UniqueName: \"kubernetes.io/projected/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-kube-api-access-hsnx6\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785820 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-combined-ca-bundle\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785862 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-db-sync-config-data\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785897 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88czv\" (UniqueName: \"kubernetes.io/projected/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-kube-api-access-88czv\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785955 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-horizon-secret-key\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.785981 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-db-sync-config-data\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786014 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlgh9\" (UniqueName: \"kubernetes.io/projected/81f50f60-4235-4365-84dd-3cfcf9142ae4-kube-api-access-tlgh9\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786056 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-combined-ca-bundle\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786093 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-scripts\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786118 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-config-data\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786147 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-logs\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786173 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-logs\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.786199 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-config\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.788477 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-config-data\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.788927 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-logs\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.789346 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-scripts\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.792286 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-combined-ca-bundle\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.792575 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-scripts\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.794315 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-db-sync-config-data\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.796958 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-horizon-secret-key\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.797719 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-combined-ca-bundle\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.800000 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-db-sync-config-data\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.801221 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-config-data\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.809998 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-combined-ca-bundle\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.819071 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-config\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.820743 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkwqs\" (UniqueName: \"kubernetes.io/projected/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-kube-api-access-bkwqs\") pod \"neutron-db-sync-psk4l\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.822148 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlgh9\" (UniqueName: \"kubernetes.io/projected/81f50f60-4235-4365-84dd-3cfcf9142ae4-kube-api-access-tlgh9\") pod \"cinder-db-sync-cqrdj\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.822168 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.825905 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.835148 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.835569 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.845807 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-9n7xj"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.849868 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88czv\" (UniqueName: \"kubernetes.io/projected/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-kube-api-access-88czv\") pod \"horizon-6c484f75c7-xh677\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.851058 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlt7q\" (UniqueName: \"kubernetes.io/projected/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-kube-api-access-xlt7q\") pod \"barbican-db-sync-2kjtg\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.869537 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-84b984fcb9-z8ww6"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.872630 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.879848 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84b984fcb9-z8ww6"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891008 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-logs\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891189 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-config-data\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891212 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-combined-ca-bundle\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891241 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-config-data\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891265 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-horizon-secret-key\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891295 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-scripts\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891376 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsnx6\" (UniqueName: \"kubernetes.io/projected/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-kube-api-access-hsnx6\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891468 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv2v6\" (UniqueName: \"kubernetes.io/projected/280a9479-3dfc-41bc-91a2-115e554179cc-kube-api-access-lv2v6\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891502 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-config-data\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891540 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-run-httpd\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891548 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-logs\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891559 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-scripts\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891637 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-logs\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891698 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891757 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891791 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-scripts\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891816 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-log-httpd\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.891841 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xmnw\" (UniqueName: \"kubernetes.io/projected/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-kube-api-access-2xmnw\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.894285 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.899910 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-combined-ca-bundle\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.901594 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-scripts\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.920201 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gwv89"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.921495 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-config-data\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.921562 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsnx6\" (UniqueName: \"kubernetes.io/projected/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-kube-api-access-hsnx6\") pod \"placement-db-sync-9n7xj\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.921856 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.933367 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gwv89"] Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.968052 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-psk4l" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.994664 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.994828 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-horizon-secret-key\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.994856 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-config-data\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.994910 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995233 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995326 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-config\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995417 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv2v6\" (UniqueName: \"kubernetes.io/projected/280a9479-3dfc-41bc-91a2-115e554179cc-kube-api-access-lv2v6\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995490 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995572 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-config-data\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995643 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-run-httpd\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.996400 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-scripts\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.995674 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-scripts\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998208 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-logs\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998267 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998327 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998360 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-scripts\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998381 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xmnw\" (UniqueName: \"kubernetes.io/projected/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-kube-api-access-2xmnw\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998412 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-log-httpd\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.998608 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm65t\" (UniqueName: \"kubernetes.io/projected/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-kube-api-access-wm65t\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:26 crc kubenswrapper[4990]: I1203 13:00:26.999004 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-logs\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:26.998009 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-run-httpd\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.000217 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-log-httpd\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.001008 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-config-data\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.002768 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-horizon-secret-key\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.006908 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-config-data\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.010427 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.010680 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-scripts\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.014145 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.015813 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv2v6\" (UniqueName: \"kubernetes.io/projected/280a9479-3dfc-41bc-91a2-115e554179cc-kube-api-access-lv2v6\") pod \"ceilometer-0\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.017711 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xmnw\" (UniqueName: \"kubernetes.io/projected/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-kube-api-access-2xmnw\") pod \"horizon-84b984fcb9-z8ww6\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.027808 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.043067 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.057771 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.160419 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9n7xj" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.161428 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm65t\" (UniqueName: \"kubernetes.io/projected/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-kube-api-access-wm65t\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.161505 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.161560 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.161590 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.161625 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-config\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.161646 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.163282 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.164383 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.164521 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.164735 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.165069 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-config\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.171084 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.187216 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm65t\" (UniqueName: \"kubernetes.io/projected/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-kube-api-access-wm65t\") pod \"dnsmasq-dns-fcfdd6f9f-gwv89\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.196425 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.247506 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mmj4c"] Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.250178 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:27 crc kubenswrapper[4990]: W1203 13:00:27.492606 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc091b50_7b0c_4922_bbf1_d203a982277b.slice/crio-2a2aad2b957c0ae0274758e6fd9d12c5c58bd5dd01d87ef030572efc14a1c6b1 WatchSource:0}: Error finding container 2a2aad2b957c0ae0274758e6fd9d12c5c58bd5dd01d87ef030572efc14a1c6b1: Status 404 returned error can't find the container with id 2a2aad2b957c0ae0274758e6fd9d12c5c58bd5dd01d87ef030572efc14a1c6b1 Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.517390 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-vc5ws"] Dec 03 13:00:27 crc kubenswrapper[4990]: W1203 13:00:27.607808 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13635c51_8af0_43cf_a0bc_bce71a531b47.slice/crio-dbab03756269fc8b66cdca0bab0e01c31aceede98b44a4e2e093e3c68ee38bc1 WatchSource:0}: Error finding container dbab03756269fc8b66cdca0bab0e01c31aceede98b44a4e2e093e3c68ee38bc1: Status 404 returned error can't find the container with id dbab03756269fc8b66cdca0bab0e01c31aceede98b44a4e2e093e3c68ee38bc1 Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.807725 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-psk4l"] Dec 03 13:00:27 crc kubenswrapper[4990]: W1203 13:00:27.809548 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f8139e2_22c8_47ce_a03d_db1d0e9e2a29.slice/crio-42cb11557bc3767e597c47981364f6e6af8ef6e363b15bea6b28118434785ab9 WatchSource:0}: Error finding container 42cb11557bc3767e597c47981364f6e6af8ef6e363b15bea6b28118434785ab9: Status 404 returned error can't find the container with id 42cb11557bc3767e597c47981364f6e6af8ef6e363b15bea6b28118434785ab9 Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.824688 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mmj4c" event={"ID":"fc091b50-7b0c-4922-bbf1-d203a982277b","Type":"ContainerStarted","Data":"2a2aad2b957c0ae0274758e6fd9d12c5c58bd5dd01d87ef030572efc14a1c6b1"} Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.830971 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerName="dnsmasq-dns" containerID="cri-o://eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff" gracePeriod=10 Dec 03 13:00:27 crc kubenswrapper[4990]: I1203 13:00:27.831443 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" event={"ID":"13635c51-8af0-43cf-a0bc-bce71a531b47","Type":"ContainerStarted","Data":"dbab03756269fc8b66cdca0bab0e01c31aceede98b44a4e2e093e3c68ee38bc1"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.286916 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2kjtg"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.464171 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cqrdj"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.517963 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c484f75c7-xh677"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.550833 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.564973 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gwv89"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.568599 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84b984fcb9-z8ww6"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.733393 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-9n7xj"] Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.794890 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.826543 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-swift-storage-0\") pod \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.826641 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-nb\") pod \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.826747 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-svc\") pod \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.826842 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lz2m\" (UniqueName: \"kubernetes.io/projected/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-kube-api-access-9lz2m\") pod \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.826927 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-config\") pod \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.826986 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-sb\") pod \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\" (UID: \"bfb6ffdc-2484-46fe-ad9d-69a60125d20f\") " Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.846953 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-kube-api-access-9lz2m" (OuterVolumeSpecName: "kube-api-access-9lz2m") pod "bfb6ffdc-2484-46fe-ad9d-69a60125d20f" (UID: "bfb6ffdc-2484-46fe-ad9d-69a60125d20f"). InnerVolumeSpecName "kube-api-access-9lz2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.859690 4990 generic.go:334] "Generic (PLEG): container finished" podID="13635c51-8af0-43cf-a0bc-bce71a531b47" containerID="f130efbe23abb0aeeab951706f11284c2b0999d3a3b71b190349c73a1a07d507" exitCode=0 Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.859807 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" event={"ID":"13635c51-8af0-43cf-a0bc-bce71a531b47","Type":"ContainerDied","Data":"f130efbe23abb0aeeab951706f11284c2b0999d3a3b71b190349c73a1a07d507"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.867313 4990 generic.go:334] "Generic (PLEG): container finished" podID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerID="eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff" exitCode=0 Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.867481 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.867942 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" event={"ID":"bfb6ffdc-2484-46fe-ad9d-69a60125d20f","Type":"ContainerDied","Data":"eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.868078 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-bntzq" event={"ID":"bfb6ffdc-2484-46fe-ad9d-69a60125d20f","Type":"ContainerDied","Data":"36791013bce3e0f80366d7486304719a4b592ea700ca369a5ba07d42fec6f313"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.868101 4990 scope.go:117] "RemoveContainer" containerID="eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.870720 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cqrdj" event={"ID":"81f50f60-4235-4365-84dd-3cfcf9142ae4","Type":"ContainerStarted","Data":"e19a25957208c439533c7ce1f549e642e48d8b2d73ba0862fc911f52d49d5f01"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.874594 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c484f75c7-xh677" event={"ID":"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec","Type":"ContainerStarted","Data":"fec8fdf0ff0ee5f621c1d15a12a03788c54a68924aebab8b0510687f7a1b8fd2"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.876013 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9n7xj" event={"ID":"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a","Type":"ContainerStarted","Data":"9da773054297e04542cf53f384575d71810d889bccdba7156fb02d028d4c5fd5"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.879693 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-psk4l" event={"ID":"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29","Type":"ContainerStarted","Data":"6ff6f94a40fc098f0cc441eb99c4845617257d31c7da670e8018ed042ce12caf"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.879826 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-psk4l" event={"ID":"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29","Type":"ContainerStarted","Data":"42cb11557bc3767e597c47981364f6e6af8ef6e363b15bea6b28118434785ab9"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.881340 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b984fcb9-z8ww6" event={"ID":"eb14dfaa-3c06-49d7-9cab-2c9940c07c17","Type":"ContainerStarted","Data":"2dcbb050fc9de8826400ab7c96d6f921acd1ec6c8616a97dd6227b9cac451545"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.890676 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2kjtg" event={"ID":"5b0890cf-e6bc-43e0-98de-f37e3aba3d60","Type":"ContainerStarted","Data":"13ca0ef4e04b9391d0ad5bbc5b71c06b308ec4dbcae0ef698acf233a15ff858c"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.898559 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"280a9479-3dfc-41bc-91a2-115e554179cc","Type":"ContainerStarted","Data":"e1d5af297c510aaf8dea9f5b2df4243912cab25b8e653870afbe507bf09c9f67"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.908407 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mmj4c" event={"ID":"fc091b50-7b0c-4922-bbf1-d203a982277b","Type":"ContainerStarted","Data":"306f6a8a159f5cb93887336a7515cebebbb3ce0e673f75dca89cd4eb92a69c0b"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.926974 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bfb6ffdc-2484-46fe-ad9d-69a60125d20f" (UID: "bfb6ffdc-2484-46fe-ad9d-69a60125d20f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.928815 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-config" (OuterVolumeSpecName: "config") pod "bfb6ffdc-2484-46fe-ad9d-69a60125d20f" (UID: "bfb6ffdc-2484-46fe-ad9d-69a60125d20f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.931588 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.931633 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lz2m\" (UniqueName: \"kubernetes.io/projected/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-kube-api-access-9lz2m\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.931649 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.938791 4990 scope.go:117] "RemoveContainer" containerID="92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.943670 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" event={"ID":"92763846-ddb6-4b6c-9ebb-8cc0606a4dca","Type":"ContainerStarted","Data":"273f0647e3e58fbdf3b662556ec4eb22f02fe0f172f8b70e73e29e03ca6b20d4"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.943809 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" event={"ID":"92763846-ddb6-4b6c-9ebb-8cc0606a4dca","Type":"ContainerStarted","Data":"234b36685d08e6ac62ad3074c42877da3b365d6ccc15eda9735845328d5e90f7"} Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.945619 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bfb6ffdc-2484-46fe-ad9d-69a60125d20f" (UID: "bfb6ffdc-2484-46fe-ad9d-69a60125d20f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.958495 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-psk4l" podStartSLOduration=2.958468339 podStartE2EDuration="2.958468339s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:00:28.903091548 +0000 UTC m=+1377.045002787" watchObservedRunningTime="2025-12-03 13:00:28.958468339 +0000 UTC m=+1377.100379588" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.976758 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mmj4c" podStartSLOduration=2.976736231 podStartE2EDuration="2.976736231s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:00:28.936864959 +0000 UTC m=+1377.078776188" watchObservedRunningTime="2025-12-03 13:00:28.976736231 +0000 UTC m=+1377.118647460" Dec 03 13:00:28 crc kubenswrapper[4990]: I1203 13:00:28.985538 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bfb6ffdc-2484-46fe-ad9d-69a60125d20f" (UID: "bfb6ffdc-2484-46fe-ad9d-69a60125d20f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.005088 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bfb6ffdc-2484-46fe-ad9d-69a60125d20f" (UID: "bfb6ffdc-2484-46fe-ad9d-69a60125d20f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.022740 4990 scope.go:117] "RemoveContainer" containerID="eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff" Dec 03 13:00:29 crc kubenswrapper[4990]: E1203 13:00:29.029805 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff\": container with ID starting with eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff not found: ID does not exist" containerID="eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.029867 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff"} err="failed to get container status \"eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff\": rpc error: code = NotFound desc = could not find container \"eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff\": container with ID starting with eace29408eb9aca43860b7e5d48dc7361dae8e40c733ac95344a789dd68bbbff not found: ID does not exist" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.029896 4990 scope.go:117] "RemoveContainer" containerID="92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390" Dec 03 13:00:29 crc kubenswrapper[4990]: E1203 13:00:29.030551 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390\": container with ID starting with 92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390 not found: ID does not exist" containerID="92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.030584 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390"} err="failed to get container status \"92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390\": rpc error: code = NotFound desc = could not find container \"92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390\": container with ID starting with 92de441d6192ec514e70be584dba05539076752d3237d743eab972f95e57b390 not found: ID does not exist" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.034956 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.035165 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.035247 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb6ffdc-2484-46fe-ad9d-69a60125d20f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.244836 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-bntzq"] Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.259032 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-bntzq"] Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.293406 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.385879 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-nb\") pod \"13635c51-8af0-43cf-a0bc-bce71a531b47\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.385998 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-sb\") pod \"13635c51-8af0-43cf-a0bc-bce71a531b47\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.386031 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-swift-storage-0\") pod \"13635c51-8af0-43cf-a0bc-bce71a531b47\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.386068 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-svc\") pod \"13635c51-8af0-43cf-a0bc-bce71a531b47\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.386120 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps2fm\" (UniqueName: \"kubernetes.io/projected/13635c51-8af0-43cf-a0bc-bce71a531b47-kube-api-access-ps2fm\") pod \"13635c51-8af0-43cf-a0bc-bce71a531b47\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.386187 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-config\") pod \"13635c51-8af0-43cf-a0bc-bce71a531b47\" (UID: \"13635c51-8af0-43cf-a0bc-bce71a531b47\") " Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.398397 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13635c51-8af0-43cf-a0bc-bce71a531b47-kube-api-access-ps2fm" (OuterVolumeSpecName: "kube-api-access-ps2fm") pod "13635c51-8af0-43cf-a0bc-bce71a531b47" (UID: "13635c51-8af0-43cf-a0bc-bce71a531b47"). InnerVolumeSpecName "kube-api-access-ps2fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.412907 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "13635c51-8af0-43cf-a0bc-bce71a531b47" (UID: "13635c51-8af0-43cf-a0bc-bce71a531b47"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:29 crc kubenswrapper[4990]: I1203 13:00:29.412919 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-config" (OuterVolumeSpecName: "config") pod "13635c51-8af0-43cf-a0bc-bce71a531b47" (UID: "13635c51-8af0-43cf-a0bc-bce71a531b47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.440993 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "13635c51-8af0-43cf-a0bc-bce71a531b47" (UID: "13635c51-8af0-43cf-a0bc-bce71a531b47"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.441598 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "13635c51-8af0-43cf-a0bc-bce71a531b47" (UID: "13635c51-8af0-43cf-a0bc-bce71a531b47"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.441940 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "13635c51-8af0-43cf-a0bc-bce71a531b47" (UID: "13635c51-8af0-43cf-a0bc-bce71a531b47"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.487954 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.487987 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.488000 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.488011 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.488023 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps2fm\" (UniqueName: \"kubernetes.io/projected/13635c51-8af0-43cf-a0bc-bce71a531b47-kube-api-access-ps2fm\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.488035 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13635c51-8af0-43cf-a0bc-bce71a531b47-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.570709 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.604047 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c484f75c7-xh677"] Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.632928 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-955b6fdd9-snf5j"] Dec 03 13:00:30 crc kubenswrapper[4990]: E1203 13:00:29.633344 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13635c51-8af0-43cf-a0bc-bce71a531b47" containerName="init" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.633356 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="13635c51-8af0-43cf-a0bc-bce71a531b47" containerName="init" Dec 03 13:00:30 crc kubenswrapper[4990]: E1203 13:00:29.633392 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerName="dnsmasq-dns" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.633400 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerName="dnsmasq-dns" Dec 03 13:00:30 crc kubenswrapper[4990]: E1203 13:00:29.633425 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerName="init" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.633430 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerName="init" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.633607 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="13635c51-8af0-43cf-a0bc-bce71a531b47" containerName="init" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.633624 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" containerName="dnsmasq-dns" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.634872 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.664642 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-955b6fdd9-snf5j"] Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.691899 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-config-data\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.691948 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmxpl\" (UniqueName: \"kubernetes.io/projected/d71af812-812a-4438-b107-fcc22f2861de-kube-api-access-dmxpl\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.691973 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d71af812-812a-4438-b107-fcc22f2861de-horizon-secret-key\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.691991 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-scripts\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.692023 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71af812-812a-4438-b107-fcc22f2861de-logs\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.796433 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmxpl\" (UniqueName: \"kubernetes.io/projected/d71af812-812a-4438-b107-fcc22f2861de-kube-api-access-dmxpl\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.804834 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d71af812-812a-4438-b107-fcc22f2861de-horizon-secret-key\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.804902 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-scripts\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.805005 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71af812-812a-4438-b107-fcc22f2861de-logs\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.805359 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-config-data\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.807054 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-config-data\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.822213 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71af812-812a-4438-b107-fcc22f2861de-logs\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.823048 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-scripts\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.836969 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d71af812-812a-4438-b107-fcc22f2861de-horizon-secret-key\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.878216 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmxpl\" (UniqueName: \"kubernetes.io/projected/d71af812-812a-4438-b107-fcc22f2861de-kube-api-access-dmxpl\") pod \"horizon-955b6fdd9-snf5j\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:29.990875 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.007991 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" event={"ID":"13635c51-8af0-43cf-a0bc-bce71a531b47","Type":"ContainerDied","Data":"dbab03756269fc8b66cdca0bab0e01c31aceede98b44a4e2e093e3c68ee38bc1"} Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.008045 4990 scope.go:117] "RemoveContainer" containerID="f130efbe23abb0aeeab951706f11284c2b0999d3a3b71b190349c73a1a07d507" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.008191 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-vc5ws" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.048864 4990 generic.go:334] "Generic (PLEG): container finished" podID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerID="273f0647e3e58fbdf3b662556ec4eb22f02fe0f172f8b70e73e29e03ca6b20d4" exitCode=0 Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.049028 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" event={"ID":"92763846-ddb6-4b6c-9ebb-8cc0606a4dca","Type":"ContainerDied","Data":"273f0647e3e58fbdf3b662556ec4eb22f02fe0f172f8b70e73e29e03ca6b20d4"} Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.049103 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" event={"ID":"92763846-ddb6-4b6c-9ebb-8cc0606a4dca","Type":"ContainerStarted","Data":"b0d6a21c407b3babae32d050e3adc1b46b76736e12fdea363bfab11a34b00a3a"} Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.105162 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-vc5ws"] Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.133233 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-vc5ws"] Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.136223 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" podStartSLOduration=4.136202066 podStartE2EDuration="4.136202066s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:00:30.110324894 +0000 UTC m=+1378.252236173" watchObservedRunningTime="2025-12-03 13:00:30.136202066 +0000 UTC m=+1378.278113295" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.283824 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13635c51-8af0-43cf-a0bc-bce71a531b47" path="/var/lib/kubelet/pods/13635c51-8af0-43cf-a0bc-bce71a531b47/volumes" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.284478 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb6ffdc-2484-46fe-ad9d-69a60125d20f" path="/var/lib/kubelet/pods/bfb6ffdc-2484-46fe-ad9d-69a60125d20f/volumes" Dec 03 13:00:30 crc kubenswrapper[4990]: I1203 13:00:30.778646 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-955b6fdd9-snf5j"] Dec 03 13:00:30 crc kubenswrapper[4990]: W1203 13:00:30.804001 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd71af812_812a_4438_b107_fcc22f2861de.slice/crio-d4c35499fa5f0671d00325008390846b63e1a6a5dc6805b0a6bcc20255f32332 WatchSource:0}: Error finding container d4c35499fa5f0671d00325008390846b63e1a6a5dc6805b0a6bcc20255f32332: Status 404 returned error can't find the container with id d4c35499fa5f0671d00325008390846b63e1a6a5dc6805b0a6bcc20255f32332 Dec 03 13:00:31 crc kubenswrapper[4990]: I1203 13:00:31.074882 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-955b6fdd9-snf5j" event={"ID":"d71af812-812a-4438-b107-fcc22f2861de","Type":"ContainerStarted","Data":"d4c35499fa5f0671d00325008390846b63e1a6a5dc6805b0a6bcc20255f32332"} Dec 03 13:00:31 crc kubenswrapper[4990]: I1203 13:00:31.076816 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:33 crc kubenswrapper[4990]: I1203 13:00:33.016584 4990 scope.go:117] "RemoveContainer" containerID="04a8b6e4fe2faa7a609fe2813c2415fc71d1c8fe1b559e9614b76bcbc7c84675" Dec 03 13:00:33 crc kubenswrapper[4990]: I1203 13:00:33.157255 4990 scope.go:117] "RemoveContainer" containerID="a06b81891d9a694411581890ab8a351ba9b12f92cd4e6bcd90af0b74ba89bbeb" Dec 03 13:00:33 crc kubenswrapper[4990]: I1203 13:00:33.228375 4990 scope.go:117] "RemoveContainer" containerID="7f4680f60c9bb8422e249f901e06b55f521143dcc19daf4dddb6f21db777f599" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.365807 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84b984fcb9-z8ww6"] Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.403239 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-679fdf78f6-4xbwz"] Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.405019 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.408276 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.411538 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-679fdf78f6-4xbwz"] Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.426754 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-combined-ca-bundle\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.426808 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-tls-certs\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.426955 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-scripts\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.427012 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-config-data\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.427066 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlfpq\" (UniqueName: \"kubernetes.io/projected/242a176b-2c0d-4067-84a2-48aff8663986-kube-api-access-nlfpq\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.427149 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/242a176b-2c0d-4067-84a2-48aff8663986-logs\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.427188 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-secret-key\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.506119 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-955b6fdd9-snf5j"] Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.512044 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f6447d884-4db4z"] Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.517912 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.521964 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f6447d884-4db4z"] Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528393 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/242a176b-2c0d-4067-84a2-48aff8663986-logs\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528478 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-secret-key\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528562 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-combined-ca-bundle\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528586 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-tls-certs\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528655 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-scripts\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528698 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-config-data\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.528728 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlfpq\" (UniqueName: \"kubernetes.io/projected/242a176b-2c0d-4067-84a2-48aff8663986-kube-api-access-nlfpq\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.530402 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/242a176b-2c0d-4067-84a2-48aff8663986-logs\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.533056 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-config-data\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.535754 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-scripts\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.538258 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-secret-key\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.542332 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-tls-certs\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.543388 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-combined-ca-bundle\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.547114 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlfpq\" (UniqueName: \"kubernetes.io/projected/242a176b-2c0d-4067-84a2-48aff8663986-kube-api-access-nlfpq\") pod \"horizon-679fdf78f6-4xbwz\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630488 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jshlz\" (UniqueName: \"kubernetes.io/projected/c626101c-4745-40ce-aba1-587cb02ea499-kube-api-access-jshlz\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630569 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c626101c-4745-40ce-aba1-587cb02ea499-scripts\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630593 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c626101c-4745-40ce-aba1-587cb02ea499-config-data\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630642 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c626101c-4745-40ce-aba1-587cb02ea499-logs\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630659 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-combined-ca-bundle\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630688 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-horizon-secret-key\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.630712 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-horizon-tls-certs\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.729223 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732038 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c626101c-4745-40ce-aba1-587cb02ea499-scripts\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732073 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c626101c-4745-40ce-aba1-587cb02ea499-config-data\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732122 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c626101c-4745-40ce-aba1-587cb02ea499-logs\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732142 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-combined-ca-bundle\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732171 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-horizon-secret-key\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732190 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-horizon-tls-certs\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.732258 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jshlz\" (UniqueName: \"kubernetes.io/projected/c626101c-4745-40ce-aba1-587cb02ea499-kube-api-access-jshlz\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.733345 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c626101c-4745-40ce-aba1-587cb02ea499-scripts\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.734572 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c626101c-4745-40ce-aba1-587cb02ea499-logs\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.738167 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-horizon-secret-key\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.738924 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-horizon-tls-certs\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.742226 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c626101c-4745-40ce-aba1-587cb02ea499-combined-ca-bundle\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.752164 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jshlz\" (UniqueName: \"kubernetes.io/projected/c626101c-4745-40ce-aba1-587cb02ea499-kube-api-access-jshlz\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.810344 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c626101c-4745-40ce-aba1-587cb02ea499-config-data\") pod \"horizon-5f6447d884-4db4z\" (UID: \"c626101c-4745-40ce-aba1-587cb02ea499\") " pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:35 crc kubenswrapper[4990]: I1203 13:00:35.837412 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:00:37 crc kubenswrapper[4990]: I1203 13:00:37.252680 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:00:37 crc kubenswrapper[4990]: I1203 13:00:37.335154 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bhcb6"] Dec 03 13:00:37 crc kubenswrapper[4990]: I1203 13:00:37.336119 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" containerID="cri-o://9cf5a63ce8640f3d8e377ccce33ba35447c2751cdaa09f8217bf45f861b64f85" gracePeriod=10 Dec 03 13:00:38 crc kubenswrapper[4990]: I1203 13:00:38.277990 4990 generic.go:334] "Generic (PLEG): container finished" podID="3186fb70-d995-467f-beef-04cd48564ecc" containerID="9cf5a63ce8640f3d8e377ccce33ba35447c2751cdaa09f8217bf45f861b64f85" exitCode=0 Dec 03 13:00:38 crc kubenswrapper[4990]: I1203 13:00:38.278039 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" event={"ID":"3186fb70-d995-467f-beef-04cd48564ecc","Type":"ContainerDied","Data":"9cf5a63ce8640f3d8e377ccce33ba35447c2751cdaa09f8217bf45f861b64f85"} Dec 03 13:00:41 crc kubenswrapper[4990]: I1203 13:00:41.312876 4990 generic.go:334] "Generic (PLEG): container finished" podID="fc091b50-7b0c-4922-bbf1-d203a982277b" containerID="306f6a8a159f5cb93887336a7515cebebbb3ce0e673f75dca89cd4eb92a69c0b" exitCode=0 Dec 03 13:00:41 crc kubenswrapper[4990]: I1203 13:00:41.312959 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mmj4c" event={"ID":"fc091b50-7b0c-4922-bbf1-d203a982277b","Type":"ContainerDied","Data":"306f6a8a159f5cb93887336a7515cebebbb3ce0e673f75dca89cd4eb92a69c0b"} Dec 03 13:00:41 crc kubenswrapper[4990]: I1203 13:00:41.545948 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:00:46 crc kubenswrapper[4990]: I1203 13:00:46.544761 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:00:48 crc kubenswrapper[4990]: E1203 13:00:48.313581 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 13:00:48 crc kubenswrapper[4990]: E1203 13:00:48.314578 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n99hdfh57dh65h5d8h5f4h5d7h597h67fh566h56bh555h667hb5h6dh5fbh7h698h57dh99hd7h685h7fh664h77h64fh59dh699h65ch74h556h8dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-88czv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6c484f75c7-xh677_openstack(ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:00:48 crc kubenswrapper[4990]: E1203 13:00:48.320527 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6c484f75c7-xh677" podUID="ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" Dec 03 13:00:51 crc kubenswrapper[4990]: I1203 13:00:51.544923 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:00:51 crc kubenswrapper[4990]: I1203 13:00:51.545698 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 13:00:56 crc kubenswrapper[4990]: I1203 13:00:56.545054 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.271386 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.272773 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n544h9ch79h58h74h666h5bhc7h574h5cbhdch7hddhc5h648hc4h9h5bchcdhf4h689h67bh589h679h647h5c5h56bh647h654h559h586h5bcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dmxpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-955b6fdd9-snf5j_openstack(d71af812-812a-4438-b107-fcc22f2861de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.305106 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.305294 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hsnx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-9n7xj_openstack(2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.306575 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-9n7xj" podUID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.349035 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-955b6fdd9-snf5j" podUID="d71af812-812a-4438-b107-fcc22f2861de" Dec 03 13:01:00 crc kubenswrapper[4990]: E1203 13:01:00.487956 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-9n7xj" podUID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.545785 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.916349 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.968792 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-combined-ca-bundle\") pod \"fc091b50-7b0c-4922-bbf1-d203a982277b\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.968912 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-scripts\") pod \"fc091b50-7b0c-4922-bbf1-d203a982277b\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.969051 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z56d9\" (UniqueName: \"kubernetes.io/projected/fc091b50-7b0c-4922-bbf1-d203a982277b-kube-api-access-z56d9\") pod \"fc091b50-7b0c-4922-bbf1-d203a982277b\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.969081 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-credential-keys\") pod \"fc091b50-7b0c-4922-bbf1-d203a982277b\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.969125 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-config-data\") pod \"fc091b50-7b0c-4922-bbf1-d203a982277b\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.969198 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-fernet-keys\") pod \"fc091b50-7b0c-4922-bbf1-d203a982277b\" (UID: \"fc091b50-7b0c-4922-bbf1-d203a982277b\") " Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.974732 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fc091b50-7b0c-4922-bbf1-d203a982277b" (UID: "fc091b50-7b0c-4922-bbf1-d203a982277b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.978970 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-scripts" (OuterVolumeSpecName: "scripts") pod "fc091b50-7b0c-4922-bbf1-d203a982277b" (UID: "fc091b50-7b0c-4922-bbf1-d203a982277b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.992524 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fc091b50-7b0c-4922-bbf1-d203a982277b" (UID: "fc091b50-7b0c-4922-bbf1-d203a982277b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.992579 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc091b50-7b0c-4922-bbf1-d203a982277b-kube-api-access-z56d9" (OuterVolumeSpecName: "kube-api-access-z56d9") pod "fc091b50-7b0c-4922-bbf1-d203a982277b" (UID: "fc091b50-7b0c-4922-bbf1-d203a982277b"). InnerVolumeSpecName "kube-api-access-z56d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.996258 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc091b50-7b0c-4922-bbf1-d203a982277b" (UID: "fc091b50-7b0c-4922-bbf1-d203a982277b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:01 crc kubenswrapper[4990]: I1203 13:01:01.999832 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-config-data" (OuterVolumeSpecName: "config-data") pod "fc091b50-7b0c-4922-bbf1-d203a982277b" (UID: "fc091b50-7b0c-4922-bbf1-d203a982277b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.071854 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z56d9\" (UniqueName: \"kubernetes.io/projected/fc091b50-7b0c-4922-bbf1-d203a982277b-kube-api-access-z56d9\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.071895 4990 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.071908 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.071920 4990 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.071931 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.071943 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc091b50-7b0c-4922-bbf1-d203a982277b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.505506 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mmj4c" event={"ID":"fc091b50-7b0c-4922-bbf1-d203a982277b","Type":"ContainerDied","Data":"2a2aad2b957c0ae0274758e6fd9d12c5c58bd5dd01d87ef030572efc14a1c6b1"} Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.505835 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a2aad2b957c0ae0274758e6fd9d12c5c58bd5dd01d87ef030572efc14a1c6b1" Dec 03 13:01:02 crc kubenswrapper[4990]: I1203 13:01:02.505904 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mmj4c" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.108569 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mmj4c"] Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.118830 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mmj4c"] Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.209873 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-p924d"] Dec 03 13:01:03 crc kubenswrapper[4990]: E1203 13:01:03.210567 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc091b50-7b0c-4922-bbf1-d203a982277b" containerName="keystone-bootstrap" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.210586 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc091b50-7b0c-4922-bbf1-d203a982277b" containerName="keystone-bootstrap" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.210887 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc091b50-7b0c-4922-bbf1-d203a982277b" containerName="keystone-bootstrap" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.211713 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.213909 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.214203 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.214254 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.214358 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.214416 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t28jn" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.218344 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p924d"] Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.291885 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-fernet-keys\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.291959 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-scripts\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.292047 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz2g2\" (UniqueName: \"kubernetes.io/projected/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-kube-api-access-mz2g2\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.292064 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-credential-keys\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.292082 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-config-data\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.292114 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-combined-ca-bundle\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.394499 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-fernet-keys\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.394565 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-scripts\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.394991 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz2g2\" (UniqueName: \"kubernetes.io/projected/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-kube-api-access-mz2g2\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.395054 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-credential-keys\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.395101 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-config-data\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.395161 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-combined-ca-bundle\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.399837 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-fernet-keys\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.400065 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-combined-ca-bundle\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.402931 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-credential-keys\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.407476 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-scripts\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.408045 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-config-data\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.424548 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz2g2\" (UniqueName: \"kubernetes.io/projected/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-kube-api-access-mz2g2\") pod \"keystone-bootstrap-p924d\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:03 crc kubenswrapper[4990]: I1203 13:01:03.571117 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:04 crc kubenswrapper[4990]: I1203 13:01:04.276097 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc091b50-7b0c-4922-bbf1-d203a982277b" path="/var/lib/kubelet/pods/fc091b50-7b0c-4922-bbf1-d203a982277b/volumes" Dec 03 13:01:06 crc kubenswrapper[4990]: I1203 13:01:06.545901 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:01:11 crc kubenswrapper[4990]: I1203 13:01:11.547003 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.513509 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.517762 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.642379 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c484f75c7-xh677" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.642447 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c484f75c7-xh677" event={"ID":"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec","Type":"ContainerDied","Data":"fec8fdf0ff0ee5f621c1d15a12a03788c54a68924aebab8b0510687f7a1b8fd2"} Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.643852 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-955b6fdd9-snf5j" event={"ID":"d71af812-812a-4438-b107-fcc22f2861de","Type":"ContainerDied","Data":"d4c35499fa5f0671d00325008390846b63e1a6a5dc6805b0a6bcc20255f32332"} Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.643944 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-955b6fdd9-snf5j" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689226 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-scripts\") pod \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689326 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmxpl\" (UniqueName: \"kubernetes.io/projected/d71af812-812a-4438-b107-fcc22f2861de-kube-api-access-dmxpl\") pod \"d71af812-812a-4438-b107-fcc22f2861de\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689412 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88czv\" (UniqueName: \"kubernetes.io/projected/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-kube-api-access-88czv\") pod \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689740 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-config-data\") pod \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689800 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-horizon-secret-key\") pod \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689838 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-scripts\") pod \"d71af812-812a-4438-b107-fcc22f2861de\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689909 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71af812-812a-4438-b107-fcc22f2861de-logs\") pod \"d71af812-812a-4438-b107-fcc22f2861de\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.689965 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d71af812-812a-4438-b107-fcc22f2861de-horizon-secret-key\") pod \"d71af812-812a-4438-b107-fcc22f2861de\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690152 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-config-data\") pod \"d71af812-812a-4438-b107-fcc22f2861de\" (UID: \"d71af812-812a-4438-b107-fcc22f2861de\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690223 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-logs\") pod \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\" (UID: \"ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec\") " Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690269 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-scripts" (OuterVolumeSpecName: "scripts") pod "d71af812-812a-4438-b107-fcc22f2861de" (UID: "d71af812-812a-4438-b107-fcc22f2861de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690360 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-config-data" (OuterVolumeSpecName: "config-data") pod "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" (UID: "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690603 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d71af812-812a-4438-b107-fcc22f2861de-logs" (OuterVolumeSpecName: "logs") pod "d71af812-812a-4438-b107-fcc22f2861de" (UID: "d71af812-812a-4438-b107-fcc22f2861de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690780 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-logs" (OuterVolumeSpecName: "logs") pod "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" (UID: "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.690991 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-scripts" (OuterVolumeSpecName: "scripts") pod "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" (UID: "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691067 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-config-data" (OuterVolumeSpecName: "config-data") pod "d71af812-812a-4438-b107-fcc22f2861de" (UID: "d71af812-812a-4438-b107-fcc22f2861de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691806 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691850 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691874 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691894 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71af812-812a-4438-b107-fcc22f2861de-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691910 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d71af812-812a-4438-b107-fcc22f2861de-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.691928 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.697091 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-kube-api-access-88czv" (OuterVolumeSpecName: "kube-api-access-88czv") pod "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" (UID: "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec"). InnerVolumeSpecName "kube-api-access-88czv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.696049 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71af812-812a-4438-b107-fcc22f2861de-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d71af812-812a-4438-b107-fcc22f2861de" (UID: "d71af812-812a-4438-b107-fcc22f2861de"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.698058 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71af812-812a-4438-b107-fcc22f2861de-kube-api-access-dmxpl" (OuterVolumeSpecName: "kube-api-access-dmxpl") pod "d71af812-812a-4438-b107-fcc22f2861de" (UID: "d71af812-812a-4438-b107-fcc22f2861de"). InnerVolumeSpecName "kube-api-access-dmxpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.699814 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" (UID: "ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.794052 4990 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.794104 4990 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d71af812-812a-4438-b107-fcc22f2861de-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.794127 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmxpl\" (UniqueName: \"kubernetes.io/projected/d71af812-812a-4438-b107-fcc22f2861de-kube-api-access-dmxpl\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:17 crc kubenswrapper[4990]: I1203 13:01:17.794144 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88czv\" (UniqueName: \"kubernetes.io/projected/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec-kube-api-access-88czv\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:18 crc kubenswrapper[4990]: I1203 13:01:18.044836 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-955b6fdd9-snf5j"] Dec 03 13:01:18 crc kubenswrapper[4990]: I1203 13:01:18.058258 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-955b6fdd9-snf5j"] Dec 03 13:01:18 crc kubenswrapper[4990]: I1203 13:01:18.079433 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c484f75c7-xh677"] Dec 03 13:01:18 crc kubenswrapper[4990]: I1203 13:01:18.091420 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6c484f75c7-xh677"] Dec 03 13:01:18 crc kubenswrapper[4990]: I1203 13:01:18.276825 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec" path="/var/lib/kubelet/pods/ba1aff8d-6f3b-420b-bcbf-e911ad4a01ec/volumes" Dec 03 13:01:18 crc kubenswrapper[4990]: I1203 13:01:18.277802 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d71af812-812a-4438-b107-fcc22f2861de" path="/var/lib/kubelet/pods/d71af812-812a-4438-b107-fcc22f2861de/volumes" Dec 03 13:01:19 crc kubenswrapper[4990]: E1203 13:01:19.663409 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 13:01:19 crc kubenswrapper[4990]: E1203 13:01:19.663874 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n548h5f8h86h647h86h569h56bhd4hdchd6h9ch54h5d8hch579h7dh9ch698h679hf5h87hb5h67fh7fh676h5d4h546h64fh55fh5dch597hccq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lv2v6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(280a9479-3dfc-41bc-91a2-115e554179cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:01:20 crc kubenswrapper[4990]: E1203 13:01:20.729789 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 13:01:20 crc kubenswrapper[4990]: E1203 13:01:20.730269 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tlgh9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-cqrdj_openstack(81f50f60-4235-4365-84dd-3cfcf9142ae4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:01:20 crc kubenswrapper[4990]: E1203 13:01:20.731574 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-cqrdj" podUID="81f50f60-4235-4365-84dd-3cfcf9142ae4" Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.797205 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.964947 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfcr7\" (UniqueName: \"kubernetes.io/projected/3186fb70-d995-467f-beef-04cd48564ecc-kube-api-access-gfcr7\") pod \"3186fb70-d995-467f-beef-04cd48564ecc\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.965134 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-dns-svc\") pod \"3186fb70-d995-467f-beef-04cd48564ecc\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.965256 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-sb\") pod \"3186fb70-d995-467f-beef-04cd48564ecc\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.965350 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-config\") pod \"3186fb70-d995-467f-beef-04cd48564ecc\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.965402 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-nb\") pod \"3186fb70-d995-467f-beef-04cd48564ecc\" (UID: \"3186fb70-d995-467f-beef-04cd48564ecc\") " Dec 03 13:01:20 crc kubenswrapper[4990]: I1203 13:01:20.988525 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3186fb70-d995-467f-beef-04cd48564ecc-kube-api-access-gfcr7" (OuterVolumeSpecName: "kube-api-access-gfcr7") pod "3186fb70-d995-467f-beef-04cd48564ecc" (UID: "3186fb70-d995-467f-beef-04cd48564ecc"). InnerVolumeSpecName "kube-api-access-gfcr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.013893 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3186fb70-d995-467f-beef-04cd48564ecc" (UID: "3186fb70-d995-467f-beef-04cd48564ecc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.014983 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3186fb70-d995-467f-beef-04cd48564ecc" (UID: "3186fb70-d995-467f-beef-04cd48564ecc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.018033 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3186fb70-d995-467f-beef-04cd48564ecc" (UID: "3186fb70-d995-467f-beef-04cd48564ecc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.026017 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-config" (OuterVolumeSpecName: "config") pod "3186fb70-d995-467f-beef-04cd48564ecc" (UID: "3186fb70-d995-467f-beef-04cd48564ecc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.068012 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.068051 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfcr7\" (UniqueName: \"kubernetes.io/projected/3186fb70-d995-467f-beef-04cd48564ecc-kube-api-access-gfcr7\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.068065 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.068074 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.068083 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3186fb70-d995-467f-beef-04cd48564ecc-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:21 crc kubenswrapper[4990]: E1203 13:01:21.233507 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 13:01:21 crc kubenswrapper[4990]: E1203 13:01:21.233649 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xlt7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-2kjtg_openstack(5b0890cf-e6bc-43e0-98de-f37e3aba3d60): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:01:21 crc kubenswrapper[4990]: E1203 13:01:21.234815 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-2kjtg" podUID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.545338 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.703446 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" event={"ID":"3186fb70-d995-467f-beef-04cd48564ecc","Type":"ContainerDied","Data":"a29a0650adbf1ef931a22cbf8dc4d894d0faa7cb9e7b520e581d0f96d9903429"} Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.703575 4990 scope.go:117] "RemoveContainer" containerID="9cf5a63ce8640f3d8e377ccce33ba35447c2751cdaa09f8217bf45f861b64f85" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.703696 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bhcb6" Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.775936 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bhcb6"] Dec 03 13:01:21 crc kubenswrapper[4990]: I1203 13:01:21.785689 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bhcb6"] Dec 03 13:01:22 crc kubenswrapper[4990]: E1203 13:01:22.241146 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-2kjtg" podUID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" Dec 03 13:01:22 crc kubenswrapper[4990]: E1203 13:01:22.255881 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-cqrdj" podUID="81f50f60-4235-4365-84dd-3cfcf9142ae4" Dec 03 13:01:22 crc kubenswrapper[4990]: I1203 13:01:22.277326 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3186fb70-d995-467f-beef-04cd48564ecc" path="/var/lib/kubelet/pods/3186fb70-d995-467f-beef-04cd48564ecc/volumes" Dec 03 13:01:22 crc kubenswrapper[4990]: I1203 13:01:22.321811 4990 scope.go:117] "RemoveContainer" containerID="669c2fd5365603e7961c038baecd37544f799c4ee0a2fe0a7b85a05baf77510f" Dec 03 13:01:22 crc kubenswrapper[4990]: I1203 13:01:22.722028 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-679fdf78f6-4xbwz"] Dec 03 13:01:22 crc kubenswrapper[4990]: I1203 13:01:22.781423 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f6447d884-4db4z"] Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.024553 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p924d"] Dec 03 13:01:23 crc kubenswrapper[4990]: W1203 13:01:23.033018 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod128f332d_3a7a_4e69_a5ec_5b4bca0d7c59.slice/crio-216ac10249da002e8e106f5fec494b0ebb74390935ec3922f0a247a2240ab812 WatchSource:0}: Error finding container 216ac10249da002e8e106f5fec494b0ebb74390935ec3922f0a247a2240ab812: Status 404 returned error can't find the container with id 216ac10249da002e8e106f5fec494b0ebb74390935ec3922f0a247a2240ab812 Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.739191 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p924d" event={"ID":"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59","Type":"ContainerStarted","Data":"aab463dca2b31d447a4e07b95e29921ba560ce57a46aeb6c25e89fd4e399c7ad"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.739553 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p924d" event={"ID":"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59","Type":"ContainerStarted","Data":"216ac10249da002e8e106f5fec494b0ebb74390935ec3922f0a247a2240ab812"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.747203 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b984fcb9-z8ww6" event={"ID":"eb14dfaa-3c06-49d7-9cab-2c9940c07c17","Type":"ContainerStarted","Data":"21d66d82abff1a5c01ab54b7afe377b1bb69797415c39c0b892d770d17290f9b"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.747249 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b984fcb9-z8ww6" event={"ID":"eb14dfaa-3c06-49d7-9cab-2c9940c07c17","Type":"ContainerStarted","Data":"03f6adc409089d14faca019ce5f2925c24e73562a2ca21efcf843b89bf921cb7"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.747370 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84b984fcb9-z8ww6" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon-log" containerID="cri-o://03f6adc409089d14faca019ce5f2925c24e73562a2ca21efcf843b89bf921cb7" gracePeriod=30 Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.747730 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84b984fcb9-z8ww6" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon" containerID="cri-o://21d66d82abff1a5c01ab54b7afe377b1bb69797415c39c0b892d770d17290f9b" gracePeriod=30 Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.751963 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f6447d884-4db4z" event={"ID":"c626101c-4745-40ce-aba1-587cb02ea499","Type":"ContainerStarted","Data":"84d49d0829073c7404f6b07136493a55dce51252c2db68a0bfc63106605daf67"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.752015 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f6447d884-4db4z" event={"ID":"c626101c-4745-40ce-aba1-587cb02ea499","Type":"ContainerStarted","Data":"eac3b952c6a611a15c10830d23ebee50314727b99fecc4a1dddaa4a2e2bdb42d"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.752027 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f6447d884-4db4z" event={"ID":"c626101c-4745-40ce-aba1-587cb02ea499","Type":"ContainerStarted","Data":"a9381e15e8084c615bf71edd0c85a188ed553f1ae1a78a28b6ab17a75ad7645e"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.764200 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"280a9479-3dfc-41bc-91a2-115e554179cc","Type":"ContainerStarted","Data":"82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.766511 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-flkbm" event={"ID":"8f5bbad8-f863-461f-a580-3989b7d2ceaa","Type":"ContainerStarted","Data":"fc33847e50961eb2b69818709bfa216afc7586e78b77349b8c8fc0d7e42959f4"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.768194 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9n7xj" event={"ID":"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a","Type":"ContainerStarted","Data":"6b207d159a2501c336acd5bbdd3a5760fe0d6d0d35132b5cb94f25ed7b5850f7"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.768547 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-p924d" podStartSLOduration=20.768532868 podStartE2EDuration="20.768532868s" podCreationTimestamp="2025-12-03 13:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:01:23.756394787 +0000 UTC m=+1431.898306036" watchObservedRunningTime="2025-12-03 13:01:23.768532868 +0000 UTC m=+1431.910444097" Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.769883 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-679fdf78f6-4xbwz" event={"ID":"242a176b-2c0d-4067-84a2-48aff8663986","Type":"ContainerStarted","Data":"b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.769918 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-679fdf78f6-4xbwz" event={"ID":"242a176b-2c0d-4067-84a2-48aff8663986","Type":"ContainerStarted","Data":"56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.769929 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-679fdf78f6-4xbwz" event={"ID":"242a176b-2c0d-4067-84a2-48aff8663986","Type":"ContainerStarted","Data":"ad9815b85abcd52d18b87a28e6293289f84abed9d357a0ced26bf3a4647432b4"} Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.791919 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-84b984fcb9-z8ww6" podStartSLOduration=5.155025763 podStartE2EDuration="57.791899074s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="2025-12-03 13:00:28.578292343 +0000 UTC m=+1376.720203572" lastFinishedPulling="2025-12-03 13:01:21.215165644 +0000 UTC m=+1429.357076883" observedRunningTime="2025-12-03 13:01:23.775211054 +0000 UTC m=+1431.917122283" watchObservedRunningTime="2025-12-03 13:01:23.791899074 +0000 UTC m=+1431.933810303" Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.812704 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f6447d884-4db4z" podStartSLOduration=48.812683991 podStartE2EDuration="48.812683991s" podCreationTimestamp="2025-12-03 13:00:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:01:23.800390618 +0000 UTC m=+1431.942301847" watchObservedRunningTime="2025-12-03 13:01:23.812683991 +0000 UTC m=+1431.954595220" Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.825630 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-679fdf78f6-4xbwz" podStartSLOduration=48.825614603 podStartE2EDuration="48.825614603s" podCreationTimestamp="2025-12-03 13:00:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:01:23.817474898 +0000 UTC m=+1431.959386127" watchObservedRunningTime="2025-12-03 13:01:23.825614603 +0000 UTC m=+1431.967525832" Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.858291 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-9n7xj" podStartSLOduration=4.010299125 podStartE2EDuration="57.858267104s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="2025-12-03 13:00:28.759035779 +0000 UTC m=+1376.900947008" lastFinishedPulling="2025-12-03 13:01:22.607003758 +0000 UTC m=+1430.748914987" observedRunningTime="2025-12-03 13:01:23.837091796 +0000 UTC m=+1431.979003025" watchObservedRunningTime="2025-12-03 13:01:23.858267104 +0000 UTC m=+1432.000178333" Dec 03 13:01:23 crc kubenswrapper[4990]: I1203 13:01:23.858792 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-flkbm" podStartSLOduration=7.988845588 podStartE2EDuration="1m34.858786507s" podCreationTimestamp="2025-12-03 12:59:49 +0000 UTC" firstStartedPulling="2025-12-03 12:59:50.528128438 +0000 UTC m=+1338.670039667" lastFinishedPulling="2025-12-03 13:01:17.398069317 +0000 UTC m=+1425.539980586" observedRunningTime="2025-12-03 13:01:23.849097822 +0000 UTC m=+1431.991009051" watchObservedRunningTime="2025-12-03 13:01:23.858786507 +0000 UTC m=+1432.000697736" Dec 03 13:01:25 crc kubenswrapper[4990]: I1203 13:01:25.730523 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:01:25 crc kubenswrapper[4990]: I1203 13:01:25.731253 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:01:25 crc kubenswrapper[4990]: I1203 13:01:25.838249 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:01:25 crc kubenswrapper[4990]: I1203 13:01:25.838338 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:01:27 crc kubenswrapper[4990]: I1203 13:01:27.198105 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:01:34 crc kubenswrapper[4990]: I1203 13:01:34.266812 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:01:35 crc kubenswrapper[4990]: I1203 13:01:35.732039 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-679fdf78f6-4xbwz" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 03 13:01:35 crc kubenswrapper[4990]: I1203 13:01:35.840470 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5f6447d884-4db4z" podUID="c626101c-4745-40ce-aba1-587cb02ea499" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 13:01:45 crc kubenswrapper[4990]: E1203 13:01:45.849991 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Dec 03 13:01:45 crc kubenswrapper[4990]: E1203 13:01:45.850635 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lv2v6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(280a9479-3dfc-41bc-91a2-115e554179cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:01:46 crc kubenswrapper[4990]: I1203 13:01:46.004418 4990 generic.go:334] "Generic (PLEG): container finished" podID="128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" containerID="aab463dca2b31d447a4e07b95e29921ba560ce57a46aeb6c25e89fd4e399c7ad" exitCode=0 Dec 03 13:01:46 crc kubenswrapper[4990]: I1203 13:01:46.004487 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p924d" event={"ID":"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59","Type":"ContainerDied","Data":"aab463dca2b31d447a4e07b95e29921ba560ce57a46aeb6c25e89fd4e399c7ad"} Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.015617 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2kjtg" event={"ID":"5b0890cf-e6bc-43e0-98de-f37e3aba3d60","Type":"ContainerStarted","Data":"5dc03a09fcb7e7ece4cd181ace64de8eec00fcd8301f60344f0143c551e50943"} Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.017369 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cqrdj" event={"ID":"81f50f60-4235-4365-84dd-3cfcf9142ae4","Type":"ContainerStarted","Data":"6ea21dd01851bbc1d1cdd0e941b8c7c9214ebf307d1e384d040cada3a134ce4a"} Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.054700 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2kjtg" podStartSLOduration=3.201888167 podStartE2EDuration="1m21.054682918s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="2025-12-03 13:00:28.3024969 +0000 UTC m=+1376.444408129" lastFinishedPulling="2025-12-03 13:01:46.155291651 +0000 UTC m=+1454.297202880" observedRunningTime="2025-12-03 13:01:47.03275777 +0000 UTC m=+1455.174669009" watchObservedRunningTime="2025-12-03 13:01:47.054682918 +0000 UTC m=+1455.196594147" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.067604 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-cqrdj" podStartSLOduration=3.414307939 podStartE2EDuration="1m21.067586548s" podCreationTimestamp="2025-12-03 13:00:26 +0000 UTC" firstStartedPulling="2025-12-03 13:00:28.503611224 +0000 UTC m=+1376.645522453" lastFinishedPulling="2025-12-03 13:01:46.156889833 +0000 UTC m=+1454.298801062" observedRunningTime="2025-12-03 13:01:47.052545632 +0000 UTC m=+1455.194456881" watchObservedRunningTime="2025-12-03 13:01:47.067586548 +0000 UTC m=+1455.209497777" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.367036 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.499789 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-scripts\") pod \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.499887 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-config-data\") pod \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.499902 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-combined-ca-bundle\") pod \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.499948 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-fernet-keys\") pod \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.499995 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-credential-keys\") pod \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.500042 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz2g2\" (UniqueName: \"kubernetes.io/projected/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-kube-api-access-mz2g2\") pod \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\" (UID: \"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59\") " Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.506569 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-kube-api-access-mz2g2" (OuterVolumeSpecName: "kube-api-access-mz2g2") pod "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" (UID: "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59"). InnerVolumeSpecName "kube-api-access-mz2g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.507610 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-scripts" (OuterVolumeSpecName: "scripts") pod "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" (UID: "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.508615 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" (UID: "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.528273 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" (UID: "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.532438 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" (UID: "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.534172 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-config-data" (OuterVolumeSpecName: "config-data") pod "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" (UID: "128f332d-3a7a-4e69-a5ec-5b4bca0d7c59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.602263 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.602297 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.602309 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.602681 4990 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.602689 4990 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.602698 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz2g2\" (UniqueName: \"kubernetes.io/projected/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59-kube-api-access-mz2g2\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.779610 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:01:47 crc kubenswrapper[4990]: I1203 13:01:47.993271 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.040731 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p924d" event={"ID":"128f332d-3a7a-4e69-a5ec-5b4bca0d7c59","Type":"ContainerDied","Data":"216ac10249da002e8e106f5fec494b0ebb74390935ec3922f0a247a2240ab812"} Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.040775 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216ac10249da002e8e106f5fec494b0ebb74390935ec3922f0a247a2240ab812" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.040894 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p924d" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.219792 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6d44745d6d-9zbl5"] Dec 03 13:01:48 crc kubenswrapper[4990]: E1203 13:01:48.220155 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" containerName="keystone-bootstrap" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.220182 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" containerName="keystone-bootstrap" Dec 03 13:01:48 crc kubenswrapper[4990]: E1203 13:01:48.220198 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="init" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.220217 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="init" Dec 03 13:01:48 crc kubenswrapper[4990]: E1203 13:01:48.220226 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.220232 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.220443 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="3186fb70-d995-467f-beef-04cd48564ecc" containerName="dnsmasq-dns" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.220709 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" containerName="keystone-bootstrap" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.221233 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.224215 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.224228 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.225295 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.225301 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.225805 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t28jn" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.225923 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.234418 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6d44745d6d-9zbl5"] Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.316901 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-scripts\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317155 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-fernet-keys\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317240 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-credential-keys\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317317 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-config-data\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317388 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-public-tls-certs\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317599 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-internal-tls-certs\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317696 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbjxc\" (UniqueName: \"kubernetes.io/projected/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-kube-api-access-zbjxc\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.317787 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-combined-ca-bundle\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.419246 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-public-tls-certs\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.419298 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-internal-tls-certs\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.419358 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbjxc\" (UniqueName: \"kubernetes.io/projected/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-kube-api-access-zbjxc\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.419997 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-combined-ca-bundle\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.420196 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-scripts\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.420220 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-fernet-keys\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.420262 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-credential-keys\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.420288 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-config-data\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.423721 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-public-tls-certs\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.427001 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-credential-keys\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.427311 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-combined-ca-bundle\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.427372 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-fernet-keys\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.427476 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-scripts\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.427574 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-internal-tls-certs\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.427612 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-config-data\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.439835 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbjxc\" (UniqueName: \"kubernetes.io/projected/6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383-kube-api-access-zbjxc\") pod \"keystone-6d44745d6d-9zbl5\" (UID: \"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383\") " pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:48 crc kubenswrapper[4990]: I1203 13:01:48.538989 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:49 crc kubenswrapper[4990]: I1203 13:01:49.116024 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6d44745d6d-9zbl5"] Dec 03 13:01:49 crc kubenswrapper[4990]: I1203 13:01:49.969613 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.062776 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6d44745d6d-9zbl5" event={"ID":"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383","Type":"ContainerStarted","Data":"a33a204fce2dceb3ebf57f6d2de82fc683e60fbdd0154b16e96e282c510382d2"} Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.063115 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6d44745d6d-9zbl5" event={"ID":"6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383","Type":"ContainerStarted","Data":"36238ee2c8c3dd1ab5dc2a6264c6202759491ecf10d90dd0fd563d38a7f2123b"} Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.064783 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.068492 4990 generic.go:334] "Generic (PLEG): container finished" podID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" containerID="6b207d159a2501c336acd5bbdd3a5760fe0d6d0d35132b5cb94f25ed7b5850f7" exitCode=0 Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.068532 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9n7xj" event={"ID":"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a","Type":"ContainerDied","Data":"6b207d159a2501c336acd5bbdd3a5760fe0d6d0d35132b5cb94f25ed7b5850f7"} Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.100957 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6d44745d6d-9zbl5" podStartSLOduration=2.100938298 podStartE2EDuration="2.100938298s" podCreationTimestamp="2025-12-03 13:01:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:01:50.08359932 +0000 UTC m=+1458.225510539" watchObservedRunningTime="2025-12-03 13:01:50.100938298 +0000 UTC m=+1458.242849527" Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.250977 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5f6447d884-4db4z" Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.308837 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-679fdf78f6-4xbwz"] Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.309296 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-679fdf78f6-4xbwz" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon-log" containerID="cri-o://56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213" gracePeriod=30 Dec 03 13:01:50 crc kubenswrapper[4990]: I1203 13:01:50.309802 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-679fdf78f6-4xbwz" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" containerID="cri-o://b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa" gracePeriod=30 Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.626210 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9n7xj" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.815218 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-logs\") pod \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.815378 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-config-data\") pod \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.815715 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-logs" (OuterVolumeSpecName: "logs") pod "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" (UID: "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.817072 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsnx6\" (UniqueName: \"kubernetes.io/projected/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-kube-api-access-hsnx6\") pod \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.817136 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-scripts\") pod \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.817162 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-combined-ca-bundle\") pod \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\" (UID: \"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a\") " Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.817572 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.821896 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-scripts" (OuterVolumeSpecName: "scripts") pod "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" (UID: "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.822605 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-kube-api-access-hsnx6" (OuterVolumeSpecName: "kube-api-access-hsnx6") pod "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" (UID: "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a"). InnerVolumeSpecName "kube-api-access-hsnx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.840584 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" (UID: "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.850758 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-config-data" (OuterVolumeSpecName: "config-data") pod "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" (UID: "2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.919296 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.919651 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsnx6\" (UniqueName: \"kubernetes.io/projected/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-kube-api-access-hsnx6\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.919695 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:52 crc kubenswrapper[4990]: I1203 13:01:52.919707 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.094654 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9n7xj" event={"ID":"2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a","Type":"ContainerDied","Data":"9da773054297e04542cf53f384575d71810d889bccdba7156fb02d028d4c5fd5"} Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.094705 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9da773054297e04542cf53f384575d71810d889bccdba7156fb02d028d4c5fd5" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.094762 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9n7xj" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.734190 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-644b9ffb4d-v4tbk"] Dec 03 13:01:53 crc kubenswrapper[4990]: E1203 13:01:53.734852 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" containerName="placement-db-sync" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.734876 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" containerName="placement-db-sync" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.735300 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" containerName="placement-db-sync" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.737025 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.739568 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.740974 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n66n7" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.747404 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.748073 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.748430 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.770868 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-644b9ffb4d-v4tbk"] Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940122 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-config-data\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940212 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-public-tls-certs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940261 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-scripts\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940335 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-combined-ca-bundle\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940391 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4ld2\" (UniqueName: \"kubernetes.io/projected/11b365c1-27dc-4c2f-a79c-12069d221f7f-kube-api-access-d4ld2\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940426 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-internal-tls-certs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:53 crc kubenswrapper[4990]: I1203 13:01:53.940538 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11b365c1-27dc-4c2f-a79c-12069d221f7f-logs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.042119 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-scripts\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.042193 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-combined-ca-bundle\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.042278 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4ld2\" (UniqueName: \"kubernetes.io/projected/11b365c1-27dc-4c2f-a79c-12069d221f7f-kube-api-access-d4ld2\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.042310 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-internal-tls-certs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.042690 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11b365c1-27dc-4c2f-a79c-12069d221f7f-logs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.043061 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11b365c1-27dc-4c2f-a79c-12069d221f7f-logs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.043845 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-config-data\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.043910 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-public-tls-certs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.046368 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-internal-tls-certs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.047077 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-combined-ca-bundle\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.047286 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-config-data\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.047548 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-scripts\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.048298 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b365c1-27dc-4c2f-a79c-12069d221f7f-public-tls-certs\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.061507 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4ld2\" (UniqueName: \"kubernetes.io/projected/11b365c1-27dc-4c2f-a79c-12069d221f7f-kube-api-access-d4ld2\") pod \"placement-644b9ffb4d-v4tbk\" (UID: \"11b365c1-27dc-4c2f-a79c-12069d221f7f\") " pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.069531 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.105660 4990 generic.go:334] "Generic (PLEG): container finished" podID="242a176b-2c0d-4067-84a2-48aff8663986" containerID="b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa" exitCode=0 Dec 03 13:01:54 crc kubenswrapper[4990]: I1203 13:01:54.105712 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-679fdf78f6-4xbwz" event={"ID":"242a176b-2c0d-4067-84a2-48aff8663986","Type":"ContainerDied","Data":"b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa"} Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.117054 4990 generic.go:334] "Generic (PLEG): container finished" podID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerID="21d66d82abff1a5c01ab54b7afe377b1bb69797415c39c0b892d770d17290f9b" exitCode=137 Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.117290 4990 generic.go:334] "Generic (PLEG): container finished" podID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerID="03f6adc409089d14faca019ce5f2925c24e73562a2ca21efcf843b89bf921cb7" exitCode=137 Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.117254 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b984fcb9-z8ww6" event={"ID":"eb14dfaa-3c06-49d7-9cab-2c9940c07c17","Type":"ContainerDied","Data":"21d66d82abff1a5c01ab54b7afe377b1bb69797415c39c0b892d770d17290f9b"} Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.117326 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b984fcb9-z8ww6" event={"ID":"eb14dfaa-3c06-49d7-9cab-2c9940c07c17","Type":"ContainerDied","Data":"03f6adc409089d14faca019ce5f2925c24e73562a2ca21efcf843b89bf921cb7"} Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.730531 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-679fdf78f6-4xbwz" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.740602 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.769895 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-logs\") pod \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.769937 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-horizon-secret-key\") pod \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.769967 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-config-data\") pod \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.769999 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-scripts\") pod \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.770083 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xmnw\" (UniqueName: \"kubernetes.io/projected/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-kube-api-access-2xmnw\") pod \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\" (UID: \"eb14dfaa-3c06-49d7-9cab-2c9940c07c17\") " Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.770464 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-logs" (OuterVolumeSpecName: "logs") pod "eb14dfaa-3c06-49d7-9cab-2c9940c07c17" (UID: "eb14dfaa-3c06-49d7-9cab-2c9940c07c17"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.775467 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "eb14dfaa-3c06-49d7-9cab-2c9940c07c17" (UID: "eb14dfaa-3c06-49d7-9cab-2c9940c07c17"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.780976 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-kube-api-access-2xmnw" (OuterVolumeSpecName: "kube-api-access-2xmnw") pod "eb14dfaa-3c06-49d7-9cab-2c9940c07c17" (UID: "eb14dfaa-3c06-49d7-9cab-2c9940c07c17"). InnerVolumeSpecName "kube-api-access-2xmnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.792275 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-scripts" (OuterVolumeSpecName: "scripts") pod "eb14dfaa-3c06-49d7-9cab-2c9940c07c17" (UID: "eb14dfaa-3c06-49d7-9cab-2c9940c07c17"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.796998 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-config-data" (OuterVolumeSpecName: "config-data") pod "eb14dfaa-3c06-49d7-9cab-2c9940c07c17" (UID: "eb14dfaa-3c06-49d7-9cab-2c9940c07c17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.871814 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xmnw\" (UniqueName: \"kubernetes.io/projected/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-kube-api-access-2xmnw\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.871849 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.871860 4990 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.871869 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:55 crc kubenswrapper[4990]: I1203 13:01:55.871878 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb14dfaa-3c06-49d7-9cab-2c9940c07c17-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:55 crc kubenswrapper[4990]: E1203 13:01:55.968577 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.004972 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-644b9ffb4d-v4tbk"] Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.132434 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b984fcb9-z8ww6" event={"ID":"eb14dfaa-3c06-49d7-9cab-2c9940c07c17","Type":"ContainerDied","Data":"2dcbb050fc9de8826400ab7c96d6f921acd1ec6c8616a97dd6227b9cac451545"} Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.132467 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b984fcb9-z8ww6" Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.132503 4990 scope.go:117] "RemoveContainer" containerID="21d66d82abff1a5c01ab54b7afe377b1bb69797415c39c0b892d770d17290f9b" Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.137515 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-644b9ffb4d-v4tbk" event={"ID":"11b365c1-27dc-4c2f-a79c-12069d221f7f","Type":"ContainerStarted","Data":"bdf442f4de4b52b542794d7c6656166e1f5b6f7383391d74f890d2d213dabc10"} Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.140255 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"280a9479-3dfc-41bc-91a2-115e554179cc","Type":"ContainerStarted","Data":"edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069"} Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.140382 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.140406 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="ceilometer-notification-agent" containerID="cri-o://82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39" gracePeriod=30 Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.140541 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="proxy-httpd" containerID="cri-o://edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069" gracePeriod=30 Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.186354 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84b984fcb9-z8ww6"] Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.194572 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-84b984fcb9-z8ww6"] Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.278427 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" path="/var/lib/kubelet/pods/eb14dfaa-3c06-49d7-9cab-2c9940c07c17/volumes" Dec 03 13:01:56 crc kubenswrapper[4990]: I1203 13:01:56.329499 4990 scope.go:117] "RemoveContainer" containerID="03f6adc409089d14faca019ce5f2925c24e73562a2ca21efcf843b89bf921cb7" Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.153277 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-644b9ffb4d-v4tbk" event={"ID":"11b365c1-27dc-4c2f-a79c-12069d221f7f","Type":"ContainerStarted","Data":"ef5b9c00df224d7f075157746e28746a6801a76682fc9ce1abd0a21fbd0fe66e"} Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.153773 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-644b9ffb4d-v4tbk" event={"ID":"11b365c1-27dc-4c2f-a79c-12069d221f7f","Type":"ContainerStarted","Data":"4bd3799cf918ba86dc6ef8bab61cb95440a8b9f0a1c0e15705d022101dd8d0d4"} Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.153854 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.153866 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.155648 4990 generic.go:334] "Generic (PLEG): container finished" podID="280a9479-3dfc-41bc-91a2-115e554179cc" containerID="edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069" exitCode=0 Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.155747 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"280a9479-3dfc-41bc-91a2-115e554179cc","Type":"ContainerDied","Data":"edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069"} Dec 03 13:01:57 crc kubenswrapper[4990]: I1203 13:01:57.189633 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-644b9ffb4d-v4tbk" podStartSLOduration=4.189608138 podStartE2EDuration="4.189608138s" podCreationTimestamp="2025-12-03 13:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:01:57.176701777 +0000 UTC m=+1465.318613016" watchObservedRunningTime="2025-12-03 13:01:57.189608138 +0000 UTC m=+1465.331519367" Dec 03 13:01:58 crc kubenswrapper[4990]: I1203 13:01:58.165543 4990 generic.go:334] "Generic (PLEG): container finished" podID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" containerID="5dc03a09fcb7e7ece4cd181ace64de8eec00fcd8301f60344f0143c551e50943" exitCode=0 Dec 03 13:01:58 crc kubenswrapper[4990]: I1203 13:01:58.165616 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2kjtg" event={"ID":"5b0890cf-e6bc-43e0-98de-f37e3aba3d60","Type":"ContainerDied","Data":"5dc03a09fcb7e7ece4cd181ace64de8eec00fcd8301f60344f0143c551e50943"} Dec 03 13:01:58 crc kubenswrapper[4990]: I1203 13:01:58.885437 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.042655 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-log-httpd\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.042986 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-combined-ca-bundle\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043021 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-scripts\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043076 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-run-httpd\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043314 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-sg-core-conf-yaml\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043372 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv2v6\" (UniqueName: \"kubernetes.io/projected/280a9479-3dfc-41bc-91a2-115e554179cc-kube-api-access-lv2v6\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043370 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043426 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-config-data\") pod \"280a9479-3dfc-41bc-91a2-115e554179cc\" (UID: \"280a9479-3dfc-41bc-91a2-115e554179cc\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.043577 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.044221 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.044243 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/280a9479-3dfc-41bc-91a2-115e554179cc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.049549 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-scripts" (OuterVolumeSpecName: "scripts") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.049627 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/280a9479-3dfc-41bc-91a2-115e554179cc-kube-api-access-lv2v6" (OuterVolumeSpecName: "kube-api-access-lv2v6") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "kube-api-access-lv2v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.050242 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.104405 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.113862 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-config-data" (OuterVolumeSpecName: "config-data") pod "280a9479-3dfc-41bc-91a2-115e554179cc" (UID: "280a9479-3dfc-41bc-91a2-115e554179cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.145018 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv2v6\" (UniqueName: \"kubernetes.io/projected/280a9479-3dfc-41bc-91a2-115e554179cc-kube-api-access-lv2v6\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.145053 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.145066 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.145074 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.145082 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/280a9479-3dfc-41bc-91a2-115e554179cc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.175500 4990 generic.go:334] "Generic (PLEG): container finished" podID="2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" containerID="6ff6f94a40fc098f0cc441eb99c4845617257d31c7da670e8018ed042ce12caf" exitCode=0 Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.175595 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-psk4l" event={"ID":"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29","Type":"ContainerDied","Data":"6ff6f94a40fc098f0cc441eb99c4845617257d31c7da670e8018ed042ce12caf"} Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.178532 4990 generic.go:334] "Generic (PLEG): container finished" podID="280a9479-3dfc-41bc-91a2-115e554179cc" containerID="82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39" exitCode=0 Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.178605 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"280a9479-3dfc-41bc-91a2-115e554179cc","Type":"ContainerDied","Data":"82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39"} Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.178630 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"280a9479-3dfc-41bc-91a2-115e554179cc","Type":"ContainerDied","Data":"e1d5af297c510aaf8dea9f5b2df4243912cab25b8e653870afbe507bf09c9f67"} Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.178640 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.178652 4990 scope.go:117] "RemoveContainer" containerID="edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.181253 4990 generic.go:334] "Generic (PLEG): container finished" podID="81f50f60-4235-4365-84dd-3cfcf9142ae4" containerID="6ea21dd01851bbc1d1cdd0e941b8c7c9214ebf307d1e384d040cada3a134ce4a" exitCode=0 Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.181388 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cqrdj" event={"ID":"81f50f60-4235-4365-84dd-3cfcf9142ae4","Type":"ContainerDied","Data":"6ea21dd01851bbc1d1cdd0e941b8c7c9214ebf307d1e384d040cada3a134ce4a"} Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.208421 4990 scope.go:117] "RemoveContainer" containerID="82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.242947 4990 scope.go:117] "RemoveContainer" containerID="edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069" Dec 03 13:01:59 crc kubenswrapper[4990]: E1203 13:01:59.243583 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069\": container with ID starting with edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069 not found: ID does not exist" containerID="edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.243620 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069"} err="failed to get container status \"edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069\": rpc error: code = NotFound desc = could not find container \"edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069\": container with ID starting with edec35c2dd283372cbfcdeeff6393721be96edf1b3b145180186ca5fd3560069 not found: ID does not exist" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.243649 4990 scope.go:117] "RemoveContainer" containerID="82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39" Dec 03 13:01:59 crc kubenswrapper[4990]: E1203 13:01:59.244123 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39\": container with ID starting with 82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39 not found: ID does not exist" containerID="82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.244160 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39"} err="failed to get container status \"82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39\": rpc error: code = NotFound desc = could not find container \"82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39\": container with ID starting with 82a6130f0f3f256414ec498e11a71bfe49e29a873adda447c2f926e58bfded39 not found: ID does not exist" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.248529 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.256087 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.272970 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:01:59 crc kubenswrapper[4990]: E1203 13:01:59.273569 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon-log" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.273655 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon-log" Dec 03 13:01:59 crc kubenswrapper[4990]: E1203 13:01:59.273711 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.273758 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon" Dec 03 13:01:59 crc kubenswrapper[4990]: E1203 13:01:59.273826 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="ceilometer-notification-agent" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.273885 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="ceilometer-notification-agent" Dec 03 13:01:59 crc kubenswrapper[4990]: E1203 13:01:59.273946 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="proxy-httpd" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.274064 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="proxy-httpd" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.274298 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="proxy-httpd" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.274358 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.274433 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" containerName="ceilometer-notification-agent" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.274515 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb14dfaa-3c06-49d7-9cab-2c9940c07c17" containerName="horizon-log" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.276100 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.278864 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.279170 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.302217 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.449947 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-run-httpd\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.450036 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-scripts\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.450069 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-config-data\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.450634 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.450709 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.450737 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-log-httpd\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.450833 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvz86\" (UniqueName: \"kubernetes.io/projected/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-kube-api-access-vvz86\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.470318 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552347 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvz86\" (UniqueName: \"kubernetes.io/projected/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-kube-api-access-vvz86\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552393 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-run-httpd\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552436 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-scripts\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552466 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-config-data\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552486 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552521 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552542 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-log-httpd\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.552983 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-log-httpd\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.553678 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-run-httpd\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.557397 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.557576 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-scripts\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.557962 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.559063 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-config-data\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.572049 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvz86\" (UniqueName: \"kubernetes.io/projected/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-kube-api-access-vvz86\") pod \"ceilometer-0\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.593164 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.653691 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-db-sync-config-data\") pod \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.653863 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlt7q\" (UniqueName: \"kubernetes.io/projected/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-kube-api-access-xlt7q\") pod \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.653911 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-combined-ca-bundle\") pod \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\" (UID: \"5b0890cf-e6bc-43e0-98de-f37e3aba3d60\") " Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.660661 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5b0890cf-e6bc-43e0-98de-f37e3aba3d60" (UID: "5b0890cf-e6bc-43e0-98de-f37e3aba3d60"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.660987 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-kube-api-access-xlt7q" (OuterVolumeSpecName: "kube-api-access-xlt7q") pod "5b0890cf-e6bc-43e0-98de-f37e3aba3d60" (UID: "5b0890cf-e6bc-43e0-98de-f37e3aba3d60"). InnerVolumeSpecName "kube-api-access-xlt7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.682883 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b0890cf-e6bc-43e0-98de-f37e3aba3d60" (UID: "5b0890cf-e6bc-43e0-98de-f37e3aba3d60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.756328 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlt7q\" (UniqueName: \"kubernetes.io/projected/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-kube-api-access-xlt7q\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.756363 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:01:59 crc kubenswrapper[4990]: I1203 13:01:59.756373 4990 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b0890cf-e6bc-43e0-98de-f37e3aba3d60-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.030984 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:00 crc kubenswrapper[4990]: W1203 13:02:00.032300 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cc2eddc_9da2_4549_bd1d_653a7656ff4a.slice/crio-72a639c71b273fcf0ddb5128fd6e228677f5d5633b4d7e6bc0cfaf05d8bc6821 WatchSource:0}: Error finding container 72a639c71b273fcf0ddb5128fd6e228677f5d5633b4d7e6bc0cfaf05d8bc6821: Status 404 returned error can't find the container with id 72a639c71b273fcf0ddb5128fd6e228677f5d5633b4d7e6bc0cfaf05d8bc6821 Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.192381 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerStarted","Data":"72a639c71b273fcf0ddb5128fd6e228677f5d5633b4d7e6bc0cfaf05d8bc6821"} Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.194629 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2kjtg" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.195202 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2kjtg" event={"ID":"5b0890cf-e6bc-43e0-98de-f37e3aba3d60","Type":"ContainerDied","Data":"13ca0ef4e04b9391d0ad5bbc5b71c06b308ec4dbcae0ef698acf233a15ff858c"} Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.195254 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13ca0ef4e04b9391d0ad5bbc5b71c06b308ec4dbcae0ef698acf233a15ff858c" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.279993 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="280a9479-3dfc-41bc-91a2-115e554179cc" path="/var/lib/kubelet/pods/280a9479-3dfc-41bc-91a2-115e554179cc/volumes" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.506606 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-67f9cc6bf5-jfjbk"] Dec 03 13:02:00 crc kubenswrapper[4990]: E1203 13:02:00.507192 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" containerName="barbican-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.507207 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" containerName="barbican-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.507373 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" containerName="barbican-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.514901 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.519177 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.519374 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.528760 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-h9hc9" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.535018 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-67f9cc6bf5-jfjbk"] Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.701309 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlbpp\" (UniqueName: \"kubernetes.io/projected/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-kube-api-access-qlbpp\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.701597 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-logs\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.701638 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-combined-ca-bundle\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.701711 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-config-data-custom\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.702223 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-config-data\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.707909 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.737293 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-psk4l" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.755627 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-68db7d64bb-tmc56"] Dec 03 13:02:00 crc kubenswrapper[4990]: E1203 13:02:00.756064 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" containerName="neutron-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.756080 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" containerName="neutron-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: E1203 13:02:00.756100 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f50f60-4235-4365-84dd-3cfcf9142ae4" containerName="cinder-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.756106 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f50f60-4235-4365-84dd-3cfcf9142ae4" containerName="cinder-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.756284 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f50f60-4235-4365-84dd-3cfcf9142ae4" containerName="cinder-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.756301 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" containerName="neutron-db-sync" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.757366 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.761753 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.796530 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68db7d64bb-tmc56"] Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.807468 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlbpp\" (UniqueName: \"kubernetes.io/projected/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-kube-api-access-qlbpp\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.807519 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-logs\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.807820 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-combined-ca-bundle\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.807881 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-config-data-custom\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.807929 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-config-data\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.809110 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-logs\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.818883 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-zt649"] Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.821094 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.837850 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-config-data\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.854529 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-zt649"] Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.881939 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-config-data-custom\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.886014 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-combined-ca-bundle\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.886322 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlbpp\" (UniqueName: \"kubernetes.io/projected/8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b-kube-api-access-qlbpp\") pod \"barbican-worker-67f9cc6bf5-jfjbk\" (UID: \"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b\") " pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910467 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-config\") pod \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910527 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-config-data\") pod \"81f50f60-4235-4365-84dd-3cfcf9142ae4\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910543 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81f50f60-4235-4365-84dd-3cfcf9142ae4-etc-machine-id\") pod \"81f50f60-4235-4365-84dd-3cfcf9142ae4\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910584 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-combined-ca-bundle\") pod \"81f50f60-4235-4365-84dd-3cfcf9142ae4\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910665 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlgh9\" (UniqueName: \"kubernetes.io/projected/81f50f60-4235-4365-84dd-3cfcf9142ae4-kube-api-access-tlgh9\") pod \"81f50f60-4235-4365-84dd-3cfcf9142ae4\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910689 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-db-sync-config-data\") pod \"81f50f60-4235-4365-84dd-3cfcf9142ae4\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910718 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkwqs\" (UniqueName: \"kubernetes.io/projected/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-kube-api-access-bkwqs\") pod \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.910747 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-combined-ca-bundle\") pod \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\" (UID: \"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.912932 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-scripts\") pod \"81f50f60-4235-4365-84dd-3cfcf9142ae4\" (UID: \"81f50f60-4235-4365-84dd-3cfcf9142ae4\") " Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.913461 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd6l5\" (UniqueName: \"kubernetes.io/projected/c31b816c-736a-4c7e-bcfa-832485e4ced9-kube-api-access-qd6l5\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.913538 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-config-data\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.913564 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-swift-storage-0\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.913664 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-combined-ca-bundle\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.913786 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-sb\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.913904 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-svc\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.914023 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-nb\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.914110 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc68490b-9eed-4344-a09f-96adddd6c7c7-logs\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.914148 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-config\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.914198 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-config-data-custom\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.914769 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkhqq\" (UniqueName: \"kubernetes.io/projected/fc68490b-9eed-4344-a09f-96adddd6c7c7-kube-api-access-lkhqq\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.915707 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81f50f60-4235-4365-84dd-3cfcf9142ae4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "81f50f60-4235-4365-84dd-3cfcf9142ae4" (UID: "81f50f60-4235-4365-84dd-3cfcf9142ae4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.917983 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "81f50f60-4235-4365-84dd-3cfcf9142ae4" (UID: "81f50f60-4235-4365-84dd-3cfcf9142ae4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.918919 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-kube-api-access-bkwqs" (OuterVolumeSpecName: "kube-api-access-bkwqs") pod "2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" (UID: "2f8139e2-22c8-47ce-a03d-db1d0e9e2a29"). InnerVolumeSpecName "kube-api-access-bkwqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.919890 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-scripts" (OuterVolumeSpecName: "scripts") pod "81f50f60-4235-4365-84dd-3cfcf9142ae4" (UID: "81f50f60-4235-4365-84dd-3cfcf9142ae4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.929622 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f50f60-4235-4365-84dd-3cfcf9142ae4-kube-api-access-tlgh9" (OuterVolumeSpecName: "kube-api-access-tlgh9") pod "81f50f60-4235-4365-84dd-3cfcf9142ae4" (UID: "81f50f60-4235-4365-84dd-3cfcf9142ae4"). InnerVolumeSpecName "kube-api-access-tlgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.972024 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.974557 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c844ff978-69nft"] Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.976153 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.978328 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.980999 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" (UID: "2f8139e2-22c8-47ce-a03d-db1d0e9e2a29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.990884 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81f50f60-4235-4365-84dd-3cfcf9142ae4" (UID: "81f50f60-4235-4365-84dd-3cfcf9142ae4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.992593 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c844ff978-69nft"] Dec 03 13:02:00 crc kubenswrapper[4990]: I1203 13:02:00.999734 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-config" (OuterVolumeSpecName: "config") pod "2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" (UID: "2f8139e2-22c8-47ce-a03d-db1d0e9e2a29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.005242 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-config-data" (OuterVolumeSpecName: "config-data") pod "81f50f60-4235-4365-84dd-3cfcf9142ae4" (UID: "81f50f60-4235-4365-84dd-3cfcf9142ae4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018209 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-sb\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018300 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data-custom\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018406 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rghk\" (UniqueName: \"kubernetes.io/projected/c6a6de82-b949-4b60-b844-1f7c872fe61f-kube-api-access-4rghk\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018484 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-svc\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018526 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-nb\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018600 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc68490b-9eed-4344-a09f-96adddd6c7c7-logs\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018660 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a6de82-b949-4b60-b844-1f7c872fe61f-logs\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018691 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-config\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018742 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-combined-ca-bundle\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018779 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-config-data-custom\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018856 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkhqq\" (UniqueName: \"kubernetes.io/projected/fc68490b-9eed-4344-a09f-96adddd6c7c7-kube-api-access-lkhqq\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018906 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.018951 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd6l5\" (UniqueName: \"kubernetes.io/projected/c31b816c-736a-4c7e-bcfa-832485e4ced9-kube-api-access-qd6l5\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019000 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-config-data\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019027 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-swift-storage-0\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019077 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-combined-ca-bundle\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019170 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019190 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019200 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019210 4990 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81f50f60-4235-4365-84dd-3cfcf9142ae4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019219 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019228 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlgh9\" (UniqueName: \"kubernetes.io/projected/81f50f60-4235-4365-84dd-3cfcf9142ae4-kube-api-access-tlgh9\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019236 4990 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81f50f60-4235-4365-84dd-3cfcf9142ae4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019245 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkwqs\" (UniqueName: \"kubernetes.io/projected/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-kube-api-access-bkwqs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.019254 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.020283 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-config\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.022225 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-sb\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.022815 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-swift-storage-0\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.022863 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc68490b-9eed-4344-a09f-96adddd6c7c7-logs\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.023331 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-svc\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.023638 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-nb\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.027695 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-config-data\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.028194 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-config-data-custom\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.030426 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc68490b-9eed-4344-a09f-96adddd6c7c7-combined-ca-bundle\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.039204 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd6l5\" (UniqueName: \"kubernetes.io/projected/c31b816c-736a-4c7e-bcfa-832485e4ced9-kube-api-access-qd6l5\") pod \"dnsmasq-dns-f4f4768cf-zt649\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.040506 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkhqq\" (UniqueName: \"kubernetes.io/projected/fc68490b-9eed-4344-a09f-96adddd6c7c7-kube-api-access-lkhqq\") pod \"barbican-keystone-listener-68db7d64bb-tmc56\" (UID: \"fc68490b-9eed-4344-a09f-96adddd6c7c7\") " pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.086694 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.121984 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rghk\" (UniqueName: \"kubernetes.io/projected/c6a6de82-b949-4b60-b844-1f7c872fe61f-kube-api-access-4rghk\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.122085 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a6de82-b949-4b60-b844-1f7c872fe61f-logs\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.122118 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-combined-ca-bundle\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.122177 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.122381 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data-custom\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.124340 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a6de82-b949-4b60-b844-1f7c872fe61f-logs\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.127941 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data-custom\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.129017 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-combined-ca-bundle\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.130079 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.145480 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rghk\" (UniqueName: \"kubernetes.io/projected/c6a6de82-b949-4b60-b844-1f7c872fe61f-kube-api-access-4rghk\") pod \"barbican-api-6c844ff978-69nft\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.215244 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cqrdj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.215255 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cqrdj" event={"ID":"81f50f60-4235-4365-84dd-3cfcf9142ae4","Type":"ContainerDied","Data":"e19a25957208c439533c7ce1f549e642e48d8b2d73ba0862fc911f52d49d5f01"} Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.215295 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e19a25957208c439533c7ce1f549e642e48d8b2d73ba0862fc911f52d49d5f01" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.221738 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-psk4l" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.224642 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-psk4l" event={"ID":"2f8139e2-22c8-47ce-a03d-db1d0e9e2a29","Type":"ContainerDied","Data":"42cb11557bc3767e597c47981364f6e6af8ef6e363b15bea6b28118434785ab9"} Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.224674 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42cb11557bc3767e597c47981364f6e6af8ef6e363b15bea6b28118434785ab9" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.245797 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerStarted","Data":"46657b8f8667997cdee8cdcf0a660fb31ecb761946ae3a86fec6783a83441751"} Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.326248 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.337853 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.409003 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-zt649"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.469147 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c88b6959-9sklj"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.470586 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.476402 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c88b6959-9sklj"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.520770 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.522233 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.525306 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.525573 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.525755 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nxxxj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.529717 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538571 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538690 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84d7c974-5413-4a19-bb32-a0baabde0602-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538718 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmbkc\" (UniqueName: \"kubernetes.io/projected/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-kube-api-access-gmbkc\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538758 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-config\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538799 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-svc\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538828 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-swift-storage-0\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538878 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-nb\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.538926 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.539004 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8xsq\" (UniqueName: \"kubernetes.io/projected/84d7c974-5413-4a19-bb32-a0baabde0602-kube-api-access-n8xsq\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.539057 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-scripts\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.539170 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.539200 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-sb\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.589095 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d4d966994-44d9g"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.590723 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.597164 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.597183 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wsxks" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.599077 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.603416 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 13:02:01 crc kubenswrapper[4990]: W1203 13:02:01.629009 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8914ad61_3a71_4ec2_bf91_b3d0f69cfa1b.slice/crio-aba586f42dc794da3ded80755c6ac0e3f39063b33b7cdf778440362a683cd8bf WatchSource:0}: Error finding container aba586f42dc794da3ded80755c6ac0e3f39063b33b7cdf778440362a683cd8bf: Status 404 returned error can't find the container with id aba586f42dc794da3ded80755c6ac0e3f39063b33b7cdf778440362a683cd8bf Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.641991 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642041 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-sb\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642138 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642207 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-httpd-config\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642236 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84d7c974-5413-4a19-bb32-a0baabde0602-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642260 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmbkc\" (UniqueName: \"kubernetes.io/projected/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-kube-api-access-gmbkc\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642294 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-config\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642334 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-svc\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642365 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-swift-storage-0\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642402 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-ovndb-tls-certs\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642427 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-nb\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642503 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642577 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8xsq\" (UniqueName: \"kubernetes.io/projected/84d7c974-5413-4a19-bb32-a0baabde0602-kube-api-access-n8xsq\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642601 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-config\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642646 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-scripts\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642671 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67gzf\" (UniqueName: \"kubernetes.io/projected/cab24acb-b936-4302-88ba-1ccf6f29d81e-kube-api-access-67gzf\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.642725 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-combined-ca-bundle\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.644730 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-swift-storage-0\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.645328 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-nb\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.646438 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-sb\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.646536 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84d7c974-5413-4a19-bb32-a0baabde0602-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.649196 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-svc\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.659833 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-config\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.666488 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-scripts\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.667621 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.667848 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.692723 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.693350 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmbkc\" (UniqueName: \"kubernetes.io/projected/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-kube-api-access-gmbkc\") pod \"dnsmasq-dns-54c88b6959-9sklj\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.701067 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.707306 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8xsq\" (UniqueName: \"kubernetes.io/projected/84d7c974-5413-4a19-bb32-a0baabde0602-kube-api-access-n8xsq\") pod \"cinder-scheduler-0\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.726939 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68db7d64bb-tmc56"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.743948 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-httpd-config\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.744012 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-ovndb-tls-certs\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.753676 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-config\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.753762 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67gzf\" (UniqueName: \"kubernetes.io/projected/cab24acb-b936-4302-88ba-1ccf6f29d81e-kube-api-access-67gzf\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.753811 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-combined-ca-bundle\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.768398 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-httpd-config\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.771697 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-combined-ca-bundle\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.772855 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-config\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.774429 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-67f9cc6bf5-jfjbk"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.782204 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-ovndb-tls-certs\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.785609 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67gzf\" (UniqueName: \"kubernetes.io/projected/cab24acb-b936-4302-88ba-1ccf6f29d81e-kube-api-access-67gzf\") pod \"neutron-7d4d966994-44d9g\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.811186 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d4d966994-44d9g"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.825255 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.842428 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c88b6959-9sklj"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.854404 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.861986 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.877628 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b59cf6f75-fh5j9"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.879201 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.884885 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b59cf6f75-fh5j9"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.935347 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.968088 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.971427 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 13:02:01 crc kubenswrapper[4990]: I1203 13:02:01.996639 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.063327 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-svc\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.063886 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlcck\" (UniqueName: \"kubernetes.io/projected/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-kube-api-access-qlcck\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.064124 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-nb\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.064334 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-config\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.064447 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-swift-storage-0\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.064609 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-sb\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.118653 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c844ff978-69nft"] Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.130201 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-zt649"] Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166218 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-config\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166298 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166321 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-swift-storage-0\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166387 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-logs\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166408 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-sb\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166477 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data-custom\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166541 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166576 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-svc\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166633 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-scripts\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166654 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166697 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlcck\" (UniqueName: \"kubernetes.io/projected/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-kube-api-access-qlcck\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166741 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwlq\" (UniqueName: \"kubernetes.io/projected/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-kube-api-access-wnwlq\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.166903 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-nb\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.168095 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-swift-storage-0\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.168787 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-nb\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.168932 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-svc\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.169012 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-config\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.171542 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-sb\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.190059 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlcck\" (UniqueName: \"kubernetes.io/projected/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-kube-api-access-qlcck\") pod \"dnsmasq-dns-7b59cf6f75-fh5j9\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.201693 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269123 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwlq\" (UniqueName: \"kubernetes.io/projected/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-kube-api-access-wnwlq\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269599 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269646 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-logs\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269693 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data-custom\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269725 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269773 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-scripts\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.269793 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.270681 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-logs\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.270741 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.294984 4990 generic.go:334] "Generic (PLEG): container finished" podID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" containerID="fc33847e50961eb2b69818709bfa216afc7586e78b77349b8c8fc0d7e42959f4" exitCode=0 Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.297216 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.301469 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" event={"ID":"fc68490b-9eed-4344-a09f-96adddd6c7c7","Type":"ContainerStarted","Data":"5ee2d5279e4f98ebfaf88fa816586db30545f06c8e108df2ff07557c626c78d9"} Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.301507 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-flkbm" event={"ID":"8f5bbad8-f863-461f-a580-3989b7d2ceaa","Type":"ContainerDied","Data":"fc33847e50961eb2b69818709bfa216afc7586e78b77349b8c8fc0d7e42959f4"} Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.302256 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-scripts\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.308684 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data-custom\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.309251 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwlq\" (UniqueName: \"kubernetes.io/projected/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-kube-api-access-wnwlq\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.309429 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" event={"ID":"c31b816c-736a-4c7e-bcfa-832485e4ced9","Type":"ContainerStarted","Data":"10a0c631fb2a72838b6e62bcab44e83d6b26598379a381013b2d735cd999c8ee"} Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.309515 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.340211 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" event={"ID":"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b","Type":"ContainerStarted","Data":"aba586f42dc794da3ded80755c6ac0e3f39063b33b7cdf778440362a683cd8bf"} Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.345635 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c844ff978-69nft" event={"ID":"c6a6de82-b949-4b60-b844-1f7c872fe61f","Type":"ContainerStarted","Data":"96c036f51441633a87eefc277fc2aed5a65ba49ea0125b21ad0b98943f005738"} Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.355377 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerStarted","Data":"34c01a8173b3088b912702fa17d492d878a741c491fe82c943b7e9d9de79b170"} Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.362285 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.643520 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:02 crc kubenswrapper[4990]: W1203 13:02:02.652769 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d7c974_5413_4a19_bb32_a0baabde0602.slice/crio-9481738f7106830fb9b12dc1da6a0b367f51c875cfbcbb2d5f00b2bbe2957f37 WatchSource:0}: Error finding container 9481738f7106830fb9b12dc1da6a0b367f51c875cfbcbb2d5f00b2bbe2957f37: Status 404 returned error can't find the container with id 9481738f7106830fb9b12dc1da6a0b367f51c875cfbcbb2d5f00b2bbe2957f37 Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.860216 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b59cf6f75-fh5j9"] Dec 03 13:02:02 crc kubenswrapper[4990]: W1203 13:02:02.875482 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1ece9aa_0b30_4ea9_820f_9cafbdf8fac6.slice/crio-116e241a8edbac0ecc052754fb46e69101092a053ccfe94e93758e21ec331ff4 WatchSource:0}: Error finding container 116e241a8edbac0ecc052754fb46e69101092a053ccfe94e93758e21ec331ff4: Status 404 returned error can't find the container with id 116e241a8edbac0ecc052754fb46e69101092a053ccfe94e93758e21ec331ff4 Dec 03 13:02:02 crc kubenswrapper[4990]: W1203 13:02:02.877195 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e92f8cc_28a1_4e63_951e_e3b2f2013ece.slice/crio-cac368627abc25338e9b34bc204defa7ccf6251b1acf283ba70c0cad47c87d99 WatchSource:0}: Error finding container cac368627abc25338e9b34bc204defa7ccf6251b1acf283ba70c0cad47c87d99: Status 404 returned error can't find the container with id cac368627abc25338e9b34bc204defa7ccf6251b1acf283ba70c0cad47c87d99 Dec 03 13:02:02 crc kubenswrapper[4990]: I1203 13:02:02.879885 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c88b6959-9sklj"] Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.002462 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.327839 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d4d966994-44d9g"] Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.376957 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4939c7c9-3cde-46d8-bc94-69a5e47d37d5","Type":"ContainerStarted","Data":"670c50eb99331103db8843c55ce0d1a52dcaa37dfba171057ea57850dc1fa264"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.379096 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84d7c974-5413-4a19-bb32-a0baabde0602","Type":"ContainerStarted","Data":"9481738f7106830fb9b12dc1da6a0b367f51c875cfbcbb2d5f00b2bbe2957f37"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.382377 4990 generic.go:334] "Generic (PLEG): container finished" podID="c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" containerID="f477dd75f6405a9416a9e7c13b7c02deb8a1c3f3a6799e1cb54786a20f1e9c1a" exitCode=0 Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.382476 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" event={"ID":"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6","Type":"ContainerDied","Data":"f477dd75f6405a9416a9e7c13b7c02deb8a1c3f3a6799e1cb54786a20f1e9c1a"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.382502 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" event={"ID":"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6","Type":"ContainerStarted","Data":"116e241a8edbac0ecc052754fb46e69101092a053ccfe94e93758e21ec331ff4"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.387514 4990 generic.go:334] "Generic (PLEG): container finished" podID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerID="60ea9e5ee7af29ad60abaa579b9596f7496da1860c41bf9adb318d75dd640e0e" exitCode=0 Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.387582 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" event={"ID":"6e92f8cc-28a1-4e63-951e-e3b2f2013ece","Type":"ContainerDied","Data":"60ea9e5ee7af29ad60abaa579b9596f7496da1860c41bf9adb318d75dd640e0e"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.387615 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" event={"ID":"6e92f8cc-28a1-4e63-951e-e3b2f2013ece","Type":"ContainerStarted","Data":"cac368627abc25338e9b34bc204defa7ccf6251b1acf283ba70c0cad47c87d99"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.405584 4990 generic.go:334] "Generic (PLEG): container finished" podID="c31b816c-736a-4c7e-bcfa-832485e4ced9" containerID="b723181e56cd59727c353265c051e96d37b4bdd9f9cad870ccab5601d21fa03b" exitCode=0 Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.405687 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" event={"ID":"c31b816c-736a-4c7e-bcfa-832485e4ced9","Type":"ContainerDied","Data":"b723181e56cd59727c353265c051e96d37b4bdd9f9cad870ccab5601d21fa03b"} Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.410683 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c844ff978-69nft" event={"ID":"c6a6de82-b949-4b60-b844-1f7c872fe61f","Type":"ContainerStarted","Data":"7964d3660976a869710c314952ac1c78a6425a238a2a27a39de137184a7ee7c7"} Dec 03 13:02:03 crc kubenswrapper[4990]: W1203 13:02:03.829530 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcab24acb_b936_4302_88ba_1ccf6f29d81e.slice/crio-cb35b7994990d10c29427bdba14365642a6dfad6b885051f0532d8a05fbc227d WatchSource:0}: Error finding container cb35b7994990d10c29427bdba14365642a6dfad6b885051f0532d8a05fbc227d: Status 404 returned error can't find the container with id cb35b7994990d10c29427bdba14365642a6dfad6b885051f0532d8a05fbc227d Dec 03 13:02:03 crc kubenswrapper[4990]: I1203 13:02:03.988403 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.128447 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-config\") pod \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.128558 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-nb\") pod \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.128762 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmbkc\" (UniqueName: \"kubernetes.io/projected/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-kube-api-access-gmbkc\") pod \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.130953 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-svc\") pod \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.131023 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-swift-storage-0\") pod \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.131089 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-sb\") pod \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\" (UID: \"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.141188 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-kube-api-access-gmbkc" (OuterVolumeSpecName: "kube-api-access-gmbkc") pod "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" (UID: "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6"). InnerVolumeSpecName "kube-api-access-gmbkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.155690 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" (UID: "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.158790 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" (UID: "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.161553 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" (UID: "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.169175 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" (UID: "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.169803 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-config" (OuterVolumeSpecName: "config") pod "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" (UID: "c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.234726 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmbkc\" (UniqueName: \"kubernetes.io/projected/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-kube-api-access-gmbkc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.234763 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.234774 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.234785 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.234794 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.234802 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.343291 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.391674 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-flkbm" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.410141 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.428766 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" event={"ID":"c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6","Type":"ContainerDied","Data":"116e241a8edbac0ecc052754fb46e69101092a053ccfe94e93758e21ec331ff4"} Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.428814 4990 scope.go:117] "RemoveContainer" containerID="f477dd75f6405a9416a9e7c13b7c02deb8a1c3f3a6799e1cb54786a20f1e9c1a" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.429117 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c88b6959-9sklj" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.433021 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-flkbm" event={"ID":"8f5bbad8-f863-461f-a580-3989b7d2ceaa","Type":"ContainerDied","Data":"e5ed9166d690f3232548c8bd8dbfb87f80b0d54e08bac856da70dda754882815"} Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.433167 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ed9166d690f3232548c8bd8dbfb87f80b0d54e08bac856da70dda754882815" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.433227 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-flkbm" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.435115 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" event={"ID":"c31b816c-736a-4c7e-bcfa-832485e4ced9","Type":"ContainerDied","Data":"10a0c631fb2a72838b6e62bcab44e83d6b26598379a381013b2d735cd999c8ee"} Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.435163 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-zt649" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.436532 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d4d966994-44d9g" event={"ID":"cab24acb-b936-4302-88ba-1ccf6f29d81e","Type":"ContainerStarted","Data":"cb35b7994990d10c29427bdba14365642a6dfad6b885051f0532d8a05fbc227d"} Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.438393 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4939c7c9-3cde-46d8-bc94-69a5e47d37d5","Type":"ContainerStarted","Data":"802b1c2deee97b91319ecc1bb4fe5ed8cb864099b27fd17d7ad669aa068e673f"} Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.445996 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-svc\") pod \"c31b816c-736a-4c7e-bcfa-832485e4ced9\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446069 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-sb\") pod \"c31b816c-736a-4c7e-bcfa-832485e4ced9\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446123 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-combined-ca-bundle\") pod \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446172 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m96\" (UniqueName: \"kubernetes.io/projected/8f5bbad8-f863-461f-a580-3989b7d2ceaa-kube-api-access-x2m96\") pod \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446195 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-db-sync-config-data\") pod \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446215 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-swift-storage-0\") pod \"c31b816c-736a-4c7e-bcfa-832485e4ced9\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446235 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd6l5\" (UniqueName: \"kubernetes.io/projected/c31b816c-736a-4c7e-bcfa-832485e4ced9-kube-api-access-qd6l5\") pod \"c31b816c-736a-4c7e-bcfa-832485e4ced9\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446273 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-config-data\") pod \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\" (UID: \"8f5bbad8-f863-461f-a580-3989b7d2ceaa\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446293 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-nb\") pod \"c31b816c-736a-4c7e-bcfa-832485e4ced9\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.446328 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-config\") pod \"c31b816c-736a-4c7e-bcfa-832485e4ced9\" (UID: \"c31b816c-736a-4c7e-bcfa-832485e4ced9\") " Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.456329 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31b816c-736a-4c7e-bcfa-832485e4ced9-kube-api-access-qd6l5" (OuterVolumeSpecName: "kube-api-access-qd6l5") pod "c31b816c-736a-4c7e-bcfa-832485e4ced9" (UID: "c31b816c-736a-4c7e-bcfa-832485e4ced9"). InnerVolumeSpecName "kube-api-access-qd6l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.463777 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8f5bbad8-f863-461f-a580-3989b7d2ceaa" (UID: "8f5bbad8-f863-461f-a580-3989b7d2ceaa"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.482578 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f5bbad8-f863-461f-a580-3989b7d2ceaa-kube-api-access-x2m96" (OuterVolumeSpecName: "kube-api-access-x2m96") pod "8f5bbad8-f863-461f-a580-3989b7d2ceaa" (UID: "8f5bbad8-f863-461f-a580-3989b7d2ceaa"). InnerVolumeSpecName "kube-api-access-x2m96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.548675 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m96\" (UniqueName: \"kubernetes.io/projected/8f5bbad8-f863-461f-a580-3989b7d2ceaa-kube-api-access-x2m96\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.551768 4990 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.551786 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd6l5\" (UniqueName: \"kubernetes.io/projected/c31b816c-736a-4c7e-bcfa-832485e4ced9-kube-api-access-qd6l5\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.560224 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c31b816c-736a-4c7e-bcfa-832485e4ced9" (UID: "c31b816c-736a-4c7e-bcfa-832485e4ced9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.587581 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c31b816c-736a-4c7e-bcfa-832485e4ced9" (UID: "c31b816c-736a-4c7e-bcfa-832485e4ced9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.588762 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c31b816c-736a-4c7e-bcfa-832485e4ced9" (UID: "c31b816c-736a-4c7e-bcfa-832485e4ced9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.592629 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c31b816c-736a-4c7e-bcfa-832485e4ced9" (UID: "c31b816c-736a-4c7e-bcfa-832485e4ced9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.609946 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c88b6959-9sklj"] Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.614989 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c88b6959-9sklj"] Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.616952 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-config" (OuterVolumeSpecName: "config") pod "c31b816c-736a-4c7e-bcfa-832485e4ced9" (UID: "c31b816c-736a-4c7e-bcfa-832485e4ced9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.655688 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.656308 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.657194 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.657280 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.657356 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31b816c-736a-4c7e-bcfa-832485e4ced9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.656122 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-config-data" (OuterVolumeSpecName: "config-data") pod "8f5bbad8-f863-461f-a580-3989b7d2ceaa" (UID: "8f5bbad8-f863-461f-a580-3989b7d2ceaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.665593 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f5bbad8-f863-461f-a580-3989b7d2ceaa" (UID: "8f5bbad8-f863-461f-a580-3989b7d2ceaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.740288 4990 scope.go:117] "RemoveContainer" containerID="b723181e56cd59727c353265c051e96d37b4bdd9f9cad870ccab5601d21fa03b" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.759116 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:04 crc kubenswrapper[4990]: I1203 13:02:04.759437 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f5bbad8-f863-461f-a580-3989b7d2ceaa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.109974 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b59cf6f75-fh5j9"] Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.213630 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qppz4"] Dec 03 13:02:05 crc kubenswrapper[4990]: E1203 13:02:05.213997 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31b816c-736a-4c7e-bcfa-832485e4ced9" containerName="init" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.214008 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31b816c-736a-4c7e-bcfa-832485e4ced9" containerName="init" Dec 03 13:02:05 crc kubenswrapper[4990]: E1203 13:02:05.214019 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" containerName="init" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.214025 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" containerName="init" Dec 03 13:02:05 crc kubenswrapper[4990]: E1203 13:02:05.214038 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" containerName="glance-db-sync" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.214046 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" containerName="glance-db-sync" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.214233 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" containerName="glance-db-sync" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.214254 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" containerName="init" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.214268 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31b816c-736a-4c7e-bcfa-832485e4ced9" containerName="init" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.215197 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.286240 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qppz4"] Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.379014 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.379476 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.379558 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.379590 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-config\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.381213 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.381374 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g25cl\" (UniqueName: \"kubernetes.io/projected/b02bd168-995c-41d6-b835-a8e4ce46b626-kube-api-access-g25cl\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.459023 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c844ff978-69nft" event={"ID":"c6a6de82-b949-4b60-b844-1f7c872fe61f","Type":"ContainerStarted","Data":"0318580cd31a503552c885a434e1c45bd62e99f1ab0c92aab327d47761ca9faa"} Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.459141 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.459154 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.491637 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.491718 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g25cl\" (UniqueName: \"kubernetes.io/projected/b02bd168-995c-41d6-b835-a8e4ce46b626-kube-api-access-g25cl\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.491786 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.491805 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.491837 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.491872 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-config\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.492977 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-config\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.493978 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.494575 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.495713 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.498646 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.521653 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-zt649"] Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.522520 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g25cl\" (UniqueName: \"kubernetes.io/projected/b02bd168-995c-41d6-b835-a8e4ce46b626-kube-api-access-g25cl\") pod \"dnsmasq-dns-6bb4fc677f-qppz4\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.553658 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-zt649"] Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.561113 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c844ff978-69nft" podStartSLOduration=5.561089946 podStartE2EDuration="5.561089946s" podCreationTimestamp="2025-12-03 13:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:05.537282328 +0000 UTC m=+1473.679193577" watchObservedRunningTime="2025-12-03 13:02:05.561089946 +0000 UTC m=+1473.703001175" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.730793 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-679fdf78f6-4xbwz" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.776395 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.979885 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.983107 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.989098 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.989644 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.990250 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gzpbp" Dec 03 13:02:05 crc kubenswrapper[4990]: I1203 13:02:05.991961 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113005 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-scripts\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113072 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-config-data\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113122 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-logs\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113156 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113207 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jk2c\" (UniqueName: \"kubernetes.io/projected/5058c5a4-2073-4b97-a35b-1663a939be92-kube-api-access-9jk2c\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113229 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.113296 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214595 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214693 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-scripts\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214732 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-config-data\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214767 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-logs\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214805 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214853 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jk2c\" (UniqueName: \"kubernetes.io/projected/5058c5a4-2073-4b97-a35b-1663a939be92-kube-api-access-9jk2c\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.214877 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.215587 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-logs\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.215888 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.218556 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.220064 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-config-data\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.220806 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.241668 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-scripts\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.251514 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jk2c\" (UniqueName: \"kubernetes.io/projected/5058c5a4-2073-4b97-a35b-1663a939be92-kube-api-access-9jk2c\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.282143 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6" path="/var/lib/kubelet/pods/c1ece9aa-0b30-4ea9-820f-9cafbdf8fac6/volumes" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.283097 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31b816c-736a-4c7e-bcfa-832485e4ced9" path="/var/lib/kubelet/pods/c31b816c-736a-4c7e-bcfa-832485e4ced9/volumes" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.324553 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.380418 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.414670 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qppz4"] Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.485674 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4939c7c9-3cde-46d8-bc94-69a5e47d37d5","Type":"ContainerStarted","Data":"3b55f008491003744719912c52cac5a14a6b5262632875c2350a6c971236ab20"} Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.485847 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api-log" containerID="cri-o://802b1c2deee97b91319ecc1bb4fe5ed8cb864099b27fd17d7ad669aa068e673f" gracePeriod=30 Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.485896 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api" containerID="cri-o://3b55f008491003744719912c52cac5a14a6b5262632875c2350a6c971236ab20" gracePeriod=30 Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.486078 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.489764 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" event={"ID":"6e92f8cc-28a1-4e63-951e-e3b2f2013ece","Type":"ContainerStarted","Data":"7e6c9fa6385250949f4b131fd5ae573ce88a8d43697327b89e2eaf988bca164e"} Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.489955 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerName="dnsmasq-dns" containerID="cri-o://7e6c9fa6385250949f4b131fd5ae573ce88a8d43697327b89e2eaf988bca164e" gracePeriod=10 Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.490060 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.503137 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d4d966994-44d9g" event={"ID":"cab24acb-b936-4302-88ba-1ccf6f29d81e","Type":"ContainerStarted","Data":"311f3a153e8a397f7a5dbb6a13f6df38c9cfab799395632057389212ab43d43a"} Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.522419 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.522394745 podStartE2EDuration="5.522394745s" podCreationTimestamp="2025-12-03 13:02:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:06.507864922 +0000 UTC m=+1474.649776161" watchObservedRunningTime="2025-12-03 13:02:06.522394745 +0000 UTC m=+1474.664305974" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.542001 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.543640 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.547075 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.563016 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.569396 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" podStartSLOduration=5.569366124 podStartE2EDuration="5.569366124s" podCreationTimestamp="2025-12-03 13:02:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:06.532602515 +0000 UTC m=+1474.674513744" watchObservedRunningTime="2025-12-03 13:02:06.569366124 +0000 UTC m=+1474.711277353" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.621721 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-scripts\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.621793 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-config-data\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.621924 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.623241 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.624597 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-logs\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.624698 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.624760 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xftkm\" (UniqueName: \"kubernetes.io/projected/39cf70df-60b5-4d0e-b495-a11fcb3fe606-kube-api-access-xftkm\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727008 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-logs\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727572 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727618 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xftkm\" (UniqueName: \"kubernetes.io/projected/39cf70df-60b5-4d0e-b495-a11fcb3fe606-kube-api-access-xftkm\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727692 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-scripts\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727734 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-config-data\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727627 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-logs\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727806 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.727860 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.728135 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.728224 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.731693 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.731780 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-config-data\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.732607 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-scripts\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.744019 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xftkm\" (UniqueName: \"kubernetes.io/projected/39cf70df-60b5-4d0e-b495-a11fcb3fe606-kube-api-access-xftkm\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.764321 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:06 crc kubenswrapper[4990]: I1203 13:02:06.881865 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:07 crc kubenswrapper[4990]: I1203 13:02:07.512978 4990 generic.go:334] "Generic (PLEG): container finished" podID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerID="7e6c9fa6385250949f4b131fd5ae573ce88a8d43697327b89e2eaf988bca164e" exitCode=0 Dec 03 13:02:07 crc kubenswrapper[4990]: I1203 13:02:07.513015 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" event={"ID":"6e92f8cc-28a1-4e63-951e-e3b2f2013ece","Type":"ContainerDied","Data":"7e6c9fa6385250949f4b131fd5ae573ce88a8d43697327b89e2eaf988bca164e"} Dec 03 13:02:07 crc kubenswrapper[4990]: I1203 13:02:07.515215 4990 generic.go:334] "Generic (PLEG): container finished" podID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerID="3b55f008491003744719912c52cac5a14a6b5262632875c2350a6c971236ab20" exitCode=0 Dec 03 13:02:07 crc kubenswrapper[4990]: I1203 13:02:07.515236 4990 generic.go:334] "Generic (PLEG): container finished" podID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerID="802b1c2deee97b91319ecc1bb4fe5ed8cb864099b27fd17d7ad669aa068e673f" exitCode=143 Dec 03 13:02:07 crc kubenswrapper[4990]: I1203 13:02:07.515252 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4939c7c9-3cde-46d8-bc94-69a5e47d37d5","Type":"ContainerDied","Data":"3b55f008491003744719912c52cac5a14a6b5262632875c2350a6c971236ab20"} Dec 03 13:02:07 crc kubenswrapper[4990]: I1203 13:02:07.515267 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4939c7c9-3cde-46d8-bc94-69a5e47d37d5","Type":"ContainerDied","Data":"802b1c2deee97b91319ecc1bb4fe5ed8cb864099b27fd17d7ad669aa068e673f"} Dec 03 13:02:08 crc kubenswrapper[4990]: W1203 13:02:08.169054 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb02bd168_995c_41d6_b835_a8e4ce46b626.slice/crio-08e9bdc1bc2675b276990d98e37799ddb6f126470e294581ce573d1ee5923b9d WatchSource:0}: Error finding container 08e9bdc1bc2675b276990d98e37799ddb6f126470e294581ce573d1ee5923b9d: Status 404 returned error can't find the container with id 08e9bdc1bc2675b276990d98e37799ddb6f126470e294581ce573d1ee5923b9d Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.594768 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" event={"ID":"b02bd168-995c-41d6-b835-a8e4ce46b626","Type":"ContainerStarted","Data":"08e9bdc1bc2675b276990d98e37799ddb6f126470e294581ce573d1ee5923b9d"} Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.626030 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" event={"ID":"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b","Type":"ContainerStarted","Data":"9d3a208b745465e1bc818d95185568667fac2b87c013a69d6d6796653ada132d"} Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.635574 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d4d966994-44d9g" event={"ID":"cab24acb-b936-4302-88ba-1ccf6f29d81e","Type":"ContainerStarted","Data":"cb6274e5fb2c79285399004898ea1e83c59b93160a26ddfe2f391cfd7b839ccc"} Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.637760 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.648041 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerStarted","Data":"095b85a5e82e9073dd19929e78528edb07f7f9ec551f9e8fd63aaf3900f8882c"} Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.650377 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84d7c974-5413-4a19-bb32-a0baabde0602","Type":"ContainerStarted","Data":"f1820ccfa3e99d47a5ad5a438deabe316bf0a5a2394bb8a85a7e9e190555ddcf"} Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.652384 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" event={"ID":"fc68490b-9eed-4344-a09f-96adddd6c7c7","Type":"ContainerStarted","Data":"4dbaa02893a0706724fd1554f58831a03bfda015d4702955e967e670eb2ea3e5"} Dec 03 13:02:08 crc kubenswrapper[4990]: I1203 13:02:08.658305 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d4d966994-44d9g" podStartSLOduration=7.658248728 podStartE2EDuration="7.658248728s" podCreationTimestamp="2025-12-03 13:02:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:08.653815792 +0000 UTC m=+1476.795727021" watchObservedRunningTime="2025-12-03 13:02:08.658248728 +0000 UTC m=+1476.800159957" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.026440 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:09 crc kubenswrapper[4990]: W1203 13:02:09.062770 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39cf70df_60b5_4d0e_b495_a11fcb3fe606.slice/crio-3910d4c43bda93e3c039311205f4762478f62238c77a01243be32eef520fc21c WatchSource:0}: Error finding container 3910d4c43bda93e3c039311205f4762478f62238c77a01243be32eef520fc21c: Status 404 returned error can't find the container with id 3910d4c43bda93e3c039311205f4762478f62238c77a01243be32eef520fc21c Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.133777 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.300568 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.321583 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-nb\") pod \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.323018 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-config\") pod \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.323264 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-sb\") pod \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.323353 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlcck\" (UniqueName: \"kubernetes.io/projected/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-kube-api-access-qlcck\") pod \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.323445 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-svc\") pod \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.323553 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-swift-storage-0\") pod \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\" (UID: \"6e92f8cc-28a1-4e63-951e-e3b2f2013ece\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.428649 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-kube-api-access-qlcck" (OuterVolumeSpecName: "kube-api-access-qlcck") pod "6e92f8cc-28a1-4e63-951e-e3b2f2013ece" (UID: "6e92f8cc-28a1-4e63-951e-e3b2f2013ece"). InnerVolumeSpecName "kube-api-access-qlcck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.457952 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-config" (OuterVolumeSpecName: "config") pod "6e92f8cc-28a1-4e63-951e-e3b2f2013ece" (UID: "6e92f8cc-28a1-4e63-951e-e3b2f2013ece"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.476408 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e92f8cc-28a1-4e63-951e-e3b2f2013ece" (UID: "6e92f8cc-28a1-4e63-951e-e3b2f2013ece"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.489371 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlcck\" (UniqueName: \"kubernetes.io/projected/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-kube-api-access-qlcck\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.489406 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.489419 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.510620 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e92f8cc-28a1-4e63-951e-e3b2f2013ece" (UID: "6e92f8cc-28a1-4e63-951e-e3b2f2013ece"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.520159 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6e92f8cc-28a1-4e63-951e-e3b2f2013ece" (UID: "6e92f8cc-28a1-4e63-951e-e3b2f2013ece"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.523222 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e92f8cc-28a1-4e63-951e-e3b2f2013ece" (UID: "6e92f8cc-28a1-4e63-951e-e3b2f2013ece"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.591426 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.591479 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.591490 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e92f8cc-28a1-4e63-951e-e3b2f2013ece-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.605441 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.706090 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" event={"ID":"8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b","Type":"ContainerStarted","Data":"51c051e335902243f94cbbf5aa339682bfe4c0377bca9bd48e207bef5a625309"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.711118 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4939c7c9-3cde-46d8-bc94-69a5e47d37d5","Type":"ContainerDied","Data":"670c50eb99331103db8843c55ce0d1a52dcaa37dfba171057ea57850dc1fa264"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.711162 4990 scope.go:117] "RemoveContainer" containerID="3b55f008491003744719912c52cac5a14a6b5262632875c2350a6c971236ab20" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.711260 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.724640 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"39cf70df-60b5-4d0e-b495-a11fcb3fe606","Type":"ContainerStarted","Data":"3910d4c43bda93e3c039311205f4762478f62238c77a01243be32eef520fc21c"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.729740 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5058c5a4-2073-4b97-a35b-1663a939be92","Type":"ContainerStarted","Data":"032d2211eda1210f7c7578dc99e3709f1ef7d6591a5f645d77d1a05bcbf0a111"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.734272 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-67f9cc6bf5-jfjbk" podStartSLOduration=6.63370477 podStartE2EDuration="9.734259942s" podCreationTimestamp="2025-12-03 13:02:00 +0000 UTC" firstStartedPulling="2025-12-03 13:02:01.664738648 +0000 UTC m=+1469.806649877" lastFinishedPulling="2025-12-03 13:02:04.76529382 +0000 UTC m=+1472.907205049" observedRunningTime="2025-12-03 13:02:09.730065892 +0000 UTC m=+1477.871977121" watchObservedRunningTime="2025-12-03 13:02:09.734259942 +0000 UTC m=+1477.876171171" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.740953 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" event={"ID":"fc68490b-9eed-4344-a09f-96adddd6c7c7","Type":"ContainerStarted","Data":"661d4c9bd3b6552c7dd980b60695d46888aaf85aff82beb54d924cb160638bd7"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.755752 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" event={"ID":"6e92f8cc-28a1-4e63-951e-e3b2f2013ece","Type":"ContainerDied","Data":"cac368627abc25338e9b34bc204defa7ccf6251b1acf283ba70c0cad47c87d99"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.755875 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b59cf6f75-fh5j9" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.765791 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-68db7d64bb-tmc56" podStartSLOduration=6.591526068 podStartE2EDuration="9.765776194s" podCreationTimestamp="2025-12-03 13:02:00 +0000 UTC" firstStartedPulling="2025-12-03 13:02:01.595997285 +0000 UTC m=+1469.737908514" lastFinishedPulling="2025-12-03 13:02:04.770247411 +0000 UTC m=+1472.912158640" observedRunningTime="2025-12-03 13:02:09.759146949 +0000 UTC m=+1477.901058188" watchObservedRunningTime="2025-12-03 13:02:09.765776194 +0000 UTC m=+1477.907687423" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.766344 4990 generic.go:334] "Generic (PLEG): container finished" podID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerID="4aee085128209fa82325b11a3cea858fffb7e789070f83cc7030c0c5eb721e7f" exitCode=0 Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.767634 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" event={"ID":"b02bd168-995c-41d6-b835-a8e4ce46b626","Type":"ContainerDied","Data":"4aee085128209fa82325b11a3cea858fffb7e789070f83cc7030c0c5eb721e7f"} Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.794789 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnwlq\" (UniqueName: \"kubernetes.io/projected/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-kube-api-access-wnwlq\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.794862 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-logs\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.794983 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data-custom\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795036 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-etc-machine-id\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795122 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795183 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-combined-ca-bundle\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795200 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-scripts\") pod \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\" (UID: \"4939c7c9-3cde-46d8-bc94-69a5e47d37d5\") " Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795366 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-logs" (OuterVolumeSpecName: "logs") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795425 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795672 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.795704 4990 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.800599 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.802741 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-scripts" (OuterVolumeSpecName: "scripts") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.807245 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-kube-api-access-wnwlq" (OuterVolumeSpecName: "kube-api-access-wnwlq") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "kube-api-access-wnwlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.826529 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b59cf6f75-fh5j9"] Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.842801 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b59cf6f75-fh5j9"] Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.848557 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.874006 4990 scope.go:117] "RemoveContainer" containerID="802b1c2deee97b91319ecc1bb4fe5ed8cb864099b27fd17d7ad669aa068e673f" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.897524 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.897559 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.897572 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnwlq\" (UniqueName: \"kubernetes.io/projected/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-kube-api-access-wnwlq\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.897586 4990 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.903691 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data" (OuterVolumeSpecName: "config-data") pod "4939c7c9-3cde-46d8-bc94-69a5e47d37d5" (UID: "4939c7c9-3cde-46d8-bc94-69a5e47d37d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.947881 4990 scope.go:117] "RemoveContainer" containerID="7e6c9fa6385250949f4b131fd5ae573ce88a8d43697327b89e2eaf988bca164e" Dec 03 13:02:09 crc kubenswrapper[4990]: I1203 13:02:09.994352 4990 scope.go:117] "RemoveContainer" containerID="60ea9e5ee7af29ad60abaa579b9596f7496da1860c41bf9adb318d75dd640e0e" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:09.999370 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4939c7c9-3cde-46d8-bc94-69a5e47d37d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.062974 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.078151 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.138205 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:10 crc kubenswrapper[4990]: E1203 13:02:10.139182 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api-log" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139204 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api-log" Dec 03 13:02:10 crc kubenswrapper[4990]: E1203 13:02:10.139218 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerName="dnsmasq-dns" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139225 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerName="dnsmasq-dns" Dec 03 13:02:10 crc kubenswrapper[4990]: E1203 13:02:10.139247 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139253 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api" Dec 03 13:02:10 crc kubenswrapper[4990]: E1203 13:02:10.139273 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerName="init" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139278 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerName="init" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139445 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" containerName="dnsmasq-dns" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139780 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api-log" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.139810 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" containerName="cinder-api" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.141115 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.146867 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.147085 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.147231 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.147237 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.282024 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4939c7c9-3cde-46d8-bc94-69a5e47d37d5" path="/var/lib/kubelet/pods/4939c7c9-3cde-46d8-bc94-69a5e47d37d5/volumes" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.283494 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e92f8cc-28a1-4e63-951e-e3b2f2013ece" path="/var/lib/kubelet/pods/6e92f8cc-28a1-4e63-951e-e3b2f2013ece/volumes" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.311058 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.311133 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312498 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-config-data-custom\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312554 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x795f\" (UniqueName: \"kubernetes.io/projected/8534a0f8-e17f-49e3-948f-475e422d43e8-kube-api-access-x795f\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312590 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8534a0f8-e17f-49e3-948f-475e422d43e8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312621 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312667 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-scripts\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312773 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8534a0f8-e17f-49e3-948f-475e422d43e8-logs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.312867 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-config-data\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416647 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-config-data\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416784 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416808 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416865 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-config-data-custom\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416900 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x795f\" (UniqueName: \"kubernetes.io/projected/8534a0f8-e17f-49e3-948f-475e422d43e8-kube-api-access-x795f\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416939 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8534a0f8-e17f-49e3-948f-475e422d43e8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416961 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.416996 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-scripts\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.417043 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8534a0f8-e17f-49e3-948f-475e422d43e8-logs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.417575 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8534a0f8-e17f-49e3-948f-475e422d43e8-logs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.420607 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8534a0f8-e17f-49e3-948f-475e422d43e8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.430524 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.437231 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-config-data\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.439187 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.439637 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-config-data-custom\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.439673 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.445150 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8534a0f8-e17f-49e3-948f-475e422d43e8-scripts\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.446597 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x795f\" (UniqueName: \"kubernetes.io/projected/8534a0f8-e17f-49e3-948f-475e422d43e8-kube-api-access-x795f\") pod \"cinder-api-0\" (UID: \"8534a0f8-e17f-49e3-948f-475e422d43e8\") " pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.600116 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.791184 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"39cf70df-60b5-4d0e-b495-a11fcb3fe606","Type":"ContainerStarted","Data":"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0"} Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.807722 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5058c5a4-2073-4b97-a35b-1663a939be92","Type":"ContainerStarted","Data":"6c82d915bfd933bc4e50c4fa30cee1368b24b74442f39b9b97874fbee0884478"} Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.811750 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84d7c974-5413-4a19-bb32-a0baabde0602","Type":"ContainerStarted","Data":"fed207cdc0655eed97533254d7521b74e91566b96ddaf69d9876ca818914488b"} Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.824136 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" event={"ID":"b02bd168-995c-41d6-b835-a8e4ce46b626","Type":"ContainerStarted","Data":"2caad78d97f33a92bfeebdb8aece85f861e9be6a5ce3f7da82c8c7bfb1928ca7"} Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.824211 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.843725 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.72843662 podStartE2EDuration="9.843707639s" podCreationTimestamp="2025-12-03 13:02:01 +0000 UTC" firstStartedPulling="2025-12-03 13:02:02.654431297 +0000 UTC m=+1470.796342526" lastFinishedPulling="2025-12-03 13:02:04.769702316 +0000 UTC m=+1472.911613545" observedRunningTime="2025-12-03 13:02:10.828173019 +0000 UTC m=+1478.970084238" watchObservedRunningTime="2025-12-03 13:02:10.843707639 +0000 UTC m=+1478.985618868" Dec 03 13:02:10 crc kubenswrapper[4990]: I1203 13:02:10.888041 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" podStartSLOduration=5.888016407 podStartE2EDuration="5.888016407s" podCreationTimestamp="2025-12-03 13:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:10.849119601 +0000 UTC m=+1478.991030830" watchObservedRunningTime="2025-12-03 13:02:10.888016407 +0000 UTC m=+1479.029927636" Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.256007 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.855608 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.907680 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"39cf70df-60b5-4d0e-b495-a11fcb3fe606","Type":"ContainerStarted","Data":"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8"} Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.912454 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.942811 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.942793073 podStartE2EDuration="6.942793073s" podCreationTimestamp="2025-12-03 13:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:11.940258826 +0000 UTC m=+1480.082170055" watchObservedRunningTime="2025-12-03 13:02:11.942793073 +0000 UTC m=+1480.084704302" Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.953569 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerStarted","Data":"8fce1a54a5833c0dbfce830cda5f4331e0adef18050661c7a4a53eb939e277d6"} Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.953869 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.969880 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5058c5a4-2073-4b97-a35b-1663a939be92","Type":"ContainerStarted","Data":"64c4e7584fdc2437246e3e58bcd6d07264acf271e6aaf719db9d7387368edf05"} Dec 03 13:02:11 crc kubenswrapper[4990]: I1203 13:02:11.984245 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8534a0f8-e17f-49e3-948f-475e422d43e8","Type":"ContainerStarted","Data":"2c72d057fe9a8200a85eb247ebff1f9560c22c2053f01067cecb282ffd961ff7"} Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.007650 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.585850828 podStartE2EDuration="13.007623152s" podCreationTimestamp="2025-12-03 13:01:59 +0000 UTC" firstStartedPulling="2025-12-03 13:02:00.034839307 +0000 UTC m=+1468.176750536" lastFinishedPulling="2025-12-03 13:02:10.456611631 +0000 UTC m=+1478.598522860" observedRunningTime="2025-12-03 13:02:11.985731464 +0000 UTC m=+1480.127642693" watchObservedRunningTime="2025-12-03 13:02:12.007623152 +0000 UTC m=+1480.149534381" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.025569 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.040389 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.040367575 podStartE2EDuration="8.040367575s" podCreationTimestamp="2025-12-03 13:02:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:12.023040128 +0000 UTC m=+1480.164951347" watchObservedRunningTime="2025-12-03 13:02:12.040367575 +0000 UTC m=+1480.182278804" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.761191 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69b4f6458f-phdfw"] Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.764075 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.769997 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.770339 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.780708 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b4f6458f-phdfw"] Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.895204 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-internal-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.895291 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmtw6\" (UniqueName: \"kubernetes.io/projected/a486143f-bd72-4292-aaa1-194fb374862a-kube-api-access-tmtw6\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.895631 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-httpd-config\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.896574 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-config\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.896614 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-ovndb-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.896696 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-public-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.896746 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-combined-ca-bundle\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:12 crc kubenswrapper[4990]: I1203 13:02:12.998772 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8534a0f8-e17f-49e3-948f-475e422d43e8","Type":"ContainerStarted","Data":"f3e568e0698af46f8247be203985ee3129ec5b40d66d88b6a22c12b1c22243ba"} Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.000532 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-httpd-config\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.004068 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-log" containerID="cri-o://6c82d915bfd933bc4e50c4fa30cee1368b24b74442f39b9b97874fbee0884478" gracePeriod=30 Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.004398 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-httpd" containerID="cri-o://64c4e7584fdc2437246e3e58bcd6d07264acf271e6aaf719db9d7387368edf05" gracePeriod=30 Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.006968 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-config\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.007010 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-ovndb-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.007158 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-public-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.007600 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-combined-ca-bundle\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.007723 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-internal-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.007776 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmtw6\" (UniqueName: \"kubernetes.io/projected/a486143f-bd72-4292-aaa1-194fb374862a-kube-api-access-tmtw6\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.023893 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-httpd-config\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.025397 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-internal-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.028673 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-ovndb-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.034067 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-public-tls-certs\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.035745 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-config\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.044829 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmtw6\" (UniqueName: \"kubernetes.io/projected/a486143f-bd72-4292-aaa1-194fb374862a-kube-api-access-tmtw6\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.047500 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a486143f-bd72-4292-aaa1-194fb374862a-combined-ca-bundle\") pod \"neutron-69b4f6458f-phdfw\" (UID: \"a486143f-bd72-4292-aaa1-194fb374862a\") " pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.093543 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.741870 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b4f6458f-phdfw"] Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.917073 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:13 crc kubenswrapper[4990]: I1203 13:02:13.932112 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.036331 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b4f6458f-phdfw" event={"ID":"a486143f-bd72-4292-aaa1-194fb374862a","Type":"ContainerStarted","Data":"3e1c6e1743c886ca2935997ce927bfe84526329f47a8982fe61bb06f8e0903bf"} Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.048978 4990 generic.go:334] "Generic (PLEG): container finished" podID="5058c5a4-2073-4b97-a35b-1663a939be92" containerID="64c4e7584fdc2437246e3e58bcd6d07264acf271e6aaf719db9d7387368edf05" exitCode=0 Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.049017 4990 generic.go:334] "Generic (PLEG): container finished" podID="5058c5a4-2073-4b97-a35b-1663a939be92" containerID="6c82d915bfd933bc4e50c4fa30cee1368b24b74442f39b9b97874fbee0884478" exitCode=143 Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.049522 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5058c5a4-2073-4b97-a35b-1663a939be92","Type":"ContainerDied","Data":"64c4e7584fdc2437246e3e58bcd6d07264acf271e6aaf719db9d7387368edf05"} Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.049808 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5058c5a4-2073-4b97-a35b-1663a939be92","Type":"ContainerDied","Data":"6c82d915bfd933bc4e50c4fa30cee1368b24b74442f39b9b97874fbee0884478"} Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.054677 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8534a0f8-e17f-49e3-948f-475e422d43e8","Type":"ContainerStarted","Data":"6103c5a6d6b27cb87625397b4366a8c1f3d93915e07120c3541bc1807e4a442e"} Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.054867 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-log" containerID="cri-o://acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0" gracePeriod=30 Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.055540 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.055605 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-httpd" containerID="cri-o://ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8" gracePeriod=30 Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.082198 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.08218007 podStartE2EDuration="4.08218007s" podCreationTimestamp="2025-12-03 13:02:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:14.079934741 +0000 UTC m=+1482.221845980" watchObservedRunningTime="2025-12-03 13:02:14.08218007 +0000 UTC m=+1482.224091299" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.354264 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.445617 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-config-data\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.446102 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-scripts\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.446141 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jk2c\" (UniqueName: \"kubernetes.io/projected/5058c5a4-2073-4b97-a35b-1663a939be92-kube-api-access-9jk2c\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.446185 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-logs\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.446296 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-httpd-run\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.446358 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-combined-ca-bundle\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.446396 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5058c5a4-2073-4b97-a35b-1663a939be92\" (UID: \"5058c5a4-2073-4b97-a35b-1663a939be92\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.454871 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.470976 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.476369 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-logs" (OuterVolumeSpecName: "logs") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.484806 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-scripts" (OuterVolumeSpecName: "scripts") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.500687 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5058c5a4-2073-4b97-a35b-1663a939be92-kube-api-access-9jk2c" (OuterVolumeSpecName: "kube-api-access-9jk2c") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "kube-api-access-9jk2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.548932 4990 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.548976 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.548986 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.548995 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jk2c\" (UniqueName: \"kubernetes.io/projected/5058c5a4-2073-4b97-a35b-1663a939be92-kube-api-access-9jk2c\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.549009 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5058c5a4-2073-4b97-a35b-1663a939be92-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.565680 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-config-data" (OuterVolumeSpecName: "config-data") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.588858 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5058c5a4-2073-4b97-a35b-1663a939be92" (UID: "5058c5a4-2073-4b97-a35b-1663a939be92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.592803 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.650182 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.650215 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.650227 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5058c5a4-2073-4b97-a35b-1663a939be92-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.880975 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955473 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-httpd-run\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955578 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-scripts\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955624 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955676 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xftkm\" (UniqueName: \"kubernetes.io/projected/39cf70df-60b5-4d0e-b495-a11fcb3fe606-kube-api-access-xftkm\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955739 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-combined-ca-bundle\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955794 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-logs\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.955910 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-config-data\") pod \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\" (UID: \"39cf70df-60b5-4d0e-b495-a11fcb3fe606\") " Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.956268 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.956574 4990 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.962935 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-logs" (OuterVolumeSpecName: "logs") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.963684 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-scripts" (OuterVolumeSpecName: "scripts") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.965270 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:02:14 crc kubenswrapper[4990]: I1203 13:02:14.967534 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39cf70df-60b5-4d0e-b495-a11fcb3fe606-kube-api-access-xftkm" (OuterVolumeSpecName: "kube-api-access-xftkm") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "kube-api-access-xftkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.079972 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.080014 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.080773 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xftkm\" (UniqueName: \"kubernetes.io/projected/39cf70df-60b5-4d0e-b495-a11fcb3fe606-kube-api-access-xftkm\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.080799 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39cf70df-60b5-4d0e-b495-a11fcb3fe606-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.092640 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.104150 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.106526 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5058c5a4-2073-4b97-a35b-1663a939be92","Type":"ContainerDied","Data":"032d2211eda1210f7c7578dc99e3709f1ef7d6591a5f645d77d1a05bcbf0a111"} Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.106584 4990 scope.go:117] "RemoveContainer" containerID="64c4e7584fdc2437246e3e58bcd6d07264acf271e6aaf719db9d7387368edf05" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.106774 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.116603 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b4f6458f-phdfw" event={"ID":"a486143f-bd72-4292-aaa1-194fb374862a","Type":"ContainerStarted","Data":"6f65f1f763fd029c7bdee308617fa40350a4c9c8b75fec477641962d2393db62"} Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.116644 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b4f6458f-phdfw" event={"ID":"a486143f-bd72-4292-aaa1-194fb374862a","Type":"ContainerStarted","Data":"3f41b9e5709b35f0727d6e18f356e4741d444ebb6c69853f72f99f6af0cb03d3"} Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.117670 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.120280 4990 generic.go:334] "Generic (PLEG): container finished" podID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerID="ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8" exitCode=0 Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.120301 4990 generic.go:334] "Generic (PLEG): container finished" podID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerID="acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0" exitCode=143 Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.120820 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.120964 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"39cf70df-60b5-4d0e-b495-a11fcb3fe606","Type":"ContainerDied","Data":"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8"} Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.120987 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"39cf70df-60b5-4d0e-b495-a11fcb3fe606","Type":"ContainerDied","Data":"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0"} Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.120996 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"39cf70df-60b5-4d0e-b495-a11fcb3fe606","Type":"ContainerDied","Data":"3910d4c43bda93e3c039311205f4762478f62238c77a01243be32eef520fc21c"} Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.128858 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-config-data" (OuterVolumeSpecName: "config-data") pod "39cf70df-60b5-4d0e-b495-a11fcb3fe606" (UID: "39cf70df-60b5-4d0e-b495-a11fcb3fe606"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.140109 4990 scope.go:117] "RemoveContainer" containerID="6c82d915bfd933bc4e50c4fa30cee1368b24b74442f39b9b97874fbee0884478" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.171792 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69b4f6458f-phdfw" podStartSLOduration=3.171764863 podStartE2EDuration="3.171764863s" podCreationTimestamp="2025-12-03 13:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:15.154479676 +0000 UTC m=+1483.296390905" watchObservedRunningTime="2025-12-03 13:02:15.171764863 +0000 UTC m=+1483.313676092" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.188695 4990 scope.go:117] "RemoveContainer" containerID="ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.197876 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.197916 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.197928 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cf70df-60b5-4d0e-b495-a11fcb3fe606-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.202506 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.232917 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.251021 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: E1203 13:02:15.253325 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-httpd" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.253360 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-httpd" Dec 03 13:02:15 crc kubenswrapper[4990]: E1203 13:02:15.253393 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-log" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.253400 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-log" Dec 03 13:02:15 crc kubenswrapper[4990]: E1203 13:02:15.253432 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-log" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.253440 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-log" Dec 03 13:02:15 crc kubenswrapper[4990]: E1203 13:02:15.253508 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-httpd" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.253515 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-httpd" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.256367 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-log" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.270172 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-log" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.270236 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" containerName="glance-httpd" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.270288 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" containerName="glance-httpd" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.274614 4990 scope.go:117] "RemoveContainer" containerID="acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.276312 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.291345 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.293810 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.336924 4990 scope.go:117] "RemoveContainer" containerID="ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8" Dec 03 13:02:15 crc kubenswrapper[4990]: E1203 13:02:15.337530 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8\": container with ID starting with ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8 not found: ID does not exist" containerID="ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.337585 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8"} err="failed to get container status \"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8\": rpc error: code = NotFound desc = could not find container \"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8\": container with ID starting with ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8 not found: ID does not exist" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.337615 4990 scope.go:117] "RemoveContainer" containerID="acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0" Dec 03 13:02:15 crc kubenswrapper[4990]: E1203 13:02:15.339902 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0\": container with ID starting with acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0 not found: ID does not exist" containerID="acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.339970 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0"} err="failed to get container status \"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0\": rpc error: code = NotFound desc = could not find container \"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0\": container with ID starting with acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0 not found: ID does not exist" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.339997 4990 scope.go:117] "RemoveContainer" containerID="ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.342688 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8"} err="failed to get container status \"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8\": rpc error: code = NotFound desc = could not find container \"ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8\": container with ID starting with ccd9a28e903a284d33251022c34f3ea9cc118257c11e29e002b151f923def9f8 not found: ID does not exist" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.342715 4990 scope.go:117] "RemoveContainer" containerID="acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.345872 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.347597 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0"} err="failed to get container status \"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0\": rpc error: code = NotFound desc = could not find container \"acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0\": container with ID starting with acfb26e1f24caf6642e5c689ce70d57fa7088eedfbcca4910b9273379b03bef0 not found: ID does not exist" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425005 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmm48\" (UniqueName: \"kubernetes.io/projected/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-kube-api-access-gmm48\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425062 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425096 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-config-data\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425164 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425209 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425409 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425466 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-logs\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.425526 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-scripts\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.464701 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.468220 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.487423 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.491336 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.496074 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.496777 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.514740 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.528640 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529210 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-logs\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529259 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-scripts\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529407 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmm48\" (UniqueName: \"kubernetes.io/projected/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-kube-api-access-gmm48\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529433 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529467 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-config-data\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529566 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529626 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.529162 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.542374 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-logs\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.543017 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.568248 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.573870 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.578929 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-config-data\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.594827 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmm48\" (UniqueName: \"kubernetes.io/projected/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-kube-api-access-gmm48\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.599238 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-scripts\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631532 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-logs\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631635 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8hzj\" (UniqueName: \"kubernetes.io/projected/1dd859cc-c778-4e79-8e61-ae6e754e21d7-kube-api-access-q8hzj\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631686 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631736 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631773 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631792 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631824 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.631904 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.643402 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65cfc74cf4-n77zl"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.645720 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.658068 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.664092 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.664319 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.676189 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65cfc74cf4-n77zl"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.731631 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-679fdf78f6-4xbwz" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.731783 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734655 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-config-data-custom\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734712 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734735 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734767 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734797 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-combined-ca-bundle\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734823 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/630b8058-02f8-4c7e-aeb7-6cca035356ed-logs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734852 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-public-tls-certs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.734883 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735198 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735626 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-logs\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735683 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8hzj\" (UniqueName: \"kubernetes.io/projected/1dd859cc-c778-4e79-8e61-ae6e754e21d7-kube-api-access-q8hzj\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735703 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbk2q\" (UniqueName: \"kubernetes.io/projected/630b8058-02f8-4c7e-aeb7-6cca035356ed-kube-api-access-nbk2q\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735733 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735758 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-config-data\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735803 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-internal-tls-certs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.735828 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.736256 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-logs\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.736831 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.742483 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.745409 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.746707 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.764177 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8hzj\" (UniqueName: \"kubernetes.io/projected/1dd859cc-c778-4e79-8e61-ae6e754e21d7-kube-api-access-q8hzj\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.770433 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.772137 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.779220 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.809985 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837196 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-internal-tls-certs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837253 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-config-data-custom\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837313 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-combined-ca-bundle\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837336 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/630b8058-02f8-4c7e-aeb7-6cca035356ed-logs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837369 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-public-tls-certs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837438 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbk2q\" (UniqueName: \"kubernetes.io/projected/630b8058-02f8-4c7e-aeb7-6cca035356ed-kube-api-access-nbk2q\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.837486 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-config-data\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.849638 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/630b8058-02f8-4c7e-aeb7-6cca035356ed-logs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.870459 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-combined-ca-bundle\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.872031 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-internal-tls-certs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.883118 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-config-data\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.883853 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-public-tls-certs\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.883904 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbk2q\" (UniqueName: \"kubernetes.io/projected/630b8058-02f8-4c7e-aeb7-6cca035356ed-kube-api-access-nbk2q\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.895117 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/630b8058-02f8-4c7e-aeb7-6cca035356ed-config-data-custom\") pod \"barbican-api-65cfc74cf4-n77zl\" (UID: \"630b8058-02f8-4c7e-aeb7-6cca035356ed\") " pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.900712 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gwv89"] Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.901026 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerName="dnsmasq-dns" containerID="cri-o://b0d6a21c407b3babae32d050e3adc1b46b76736e12fdea363bfab11a34b00a3a" gracePeriod=10 Dec 03 13:02:15 crc kubenswrapper[4990]: I1203 13:02:15.946860 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.014132 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:16 crc kubenswrapper[4990]: E1203 13:02:16.114937 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92763846_ddb6_4b6c_9ebb_8cc0606a4dca.slice/crio-b0d6a21c407b3babae32d050e3adc1b46b76736e12fdea363bfab11a34b00a3a.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.170865 4990 generic.go:334] "Generic (PLEG): container finished" podID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerID="b0d6a21c407b3babae32d050e3adc1b46b76736e12fdea363bfab11a34b00a3a" exitCode=0 Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.171911 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" event={"ID":"92763846-ddb6-4b6c-9ebb-8cc0606a4dca","Type":"ContainerDied","Data":"b0d6a21c407b3babae32d050e3adc1b46b76736e12fdea363bfab11a34b00a3a"} Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.374062 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39cf70df-60b5-4d0e-b495-a11fcb3fe606" path="/var/lib/kubelet/pods/39cf70df-60b5-4d0e-b495-a11fcb3fe606/volumes" Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.375217 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5058c5a4-2073-4b97-a35b-1663a939be92" path="/var/lib/kubelet/pods/5058c5a4-2073-4b97-a35b-1663a939be92/volumes" Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.697928 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.801287 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.933777 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-config\") pod \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.934473 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-swift-storage-0\") pod \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.934567 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-sb\") pod \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.934604 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-svc\") pod \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.939015 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm65t\" (UniqueName: \"kubernetes.io/projected/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-kube-api-access-wm65t\") pod \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.939279 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-nb\") pod \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\" (UID: \"92763846-ddb6-4b6c-9ebb-8cc0606a4dca\") " Dec 03 13:02:16 crc kubenswrapper[4990]: I1203 13:02:16.957807 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-kube-api-access-wm65t" (OuterVolumeSpecName: "kube-api-access-wm65t") pod "92763846-ddb6-4b6c-9ebb-8cc0606a4dca" (UID: "92763846-ddb6-4b6c-9ebb-8cc0606a4dca"). InnerVolumeSpecName "kube-api-access-wm65t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.023312 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65cfc74cf4-n77zl"] Dec 03 13:02:17 crc kubenswrapper[4990]: W1203 13:02:17.028071 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod630b8058_02f8_4c7e_aeb7_6cca035356ed.slice/crio-744cab56c5c5efa868649b6c2546242ce2d3b34adf129d83ee698e103604d65b WatchSource:0}: Error finding container 744cab56c5c5efa868649b6c2546242ce2d3b34adf129d83ee698e103604d65b: Status 404 returned error can't find the container with id 744cab56c5c5efa868649b6c2546242ce2d3b34adf129d83ee698e103604d65b Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.029647 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-config" (OuterVolumeSpecName: "config") pod "92763846-ddb6-4b6c-9ebb-8cc0606a4dca" (UID: "92763846-ddb6-4b6c-9ebb-8cc0606a4dca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.048578 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm65t\" (UniqueName: \"kubernetes.io/projected/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-kube-api-access-wm65t\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.048607 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.049643 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "92763846-ddb6-4b6c-9ebb-8cc0606a4dca" (UID: "92763846-ddb6-4b6c-9ebb-8cc0606a4dca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.071027 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "92763846-ddb6-4b6c-9ebb-8cc0606a4dca" (UID: "92763846-ddb6-4b6c-9ebb-8cc0606a4dca"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.077626 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:17 crc kubenswrapper[4990]: W1203 13:02:17.112127 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37eba15b_49c1_4b0c_b6b7_8dcb613648fd.slice/crio-2e248d217af04e0c903894617f3c1ce482d504bf21791be1a0806310ef542cb1 WatchSource:0}: Error finding container 2e248d217af04e0c903894617f3c1ce482d504bf21791be1a0806310ef542cb1: Status 404 returned error can't find the container with id 2e248d217af04e0c903894617f3c1ce482d504bf21791be1a0806310ef542cb1 Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.116415 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92763846-ddb6-4b6c-9ebb-8cc0606a4dca" (UID: "92763846-ddb6-4b6c-9ebb-8cc0606a4dca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.131309 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92763846-ddb6-4b6c-9ebb-8cc0606a4dca" (UID: "92763846-ddb6-4b6c-9ebb-8cc0606a4dca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.145856 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.150919 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.150957 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.150970 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.150981 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92763846-ddb6-4b6c-9ebb-8cc0606a4dca-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.212884 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" event={"ID":"92763846-ddb6-4b6c-9ebb-8cc0606a4dca","Type":"ContainerDied","Data":"234b36685d08e6ac62ad3074c42877da3b365d6ccc15eda9735845328d5e90f7"} Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.212939 4990 scope.go:117] "RemoveContainer" containerID="b0d6a21c407b3babae32d050e3adc1b46b76736e12fdea363bfab11a34b00a3a" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.213069 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gwv89" Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.224214 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1dd859cc-c778-4e79-8e61-ae6e754e21d7","Type":"ContainerStarted","Data":"25ee23f1ad624163ccb404009371ec317fb793c1dd6e2725148bd4fca2ed622e"} Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.227983 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65cfc74cf4-n77zl" event={"ID":"630b8058-02f8-4c7e-aeb7-6cca035356ed","Type":"ContainerStarted","Data":"744cab56c5c5efa868649b6c2546242ce2d3b34adf129d83ee698e103604d65b"} Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.269979 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.280592 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37eba15b-49c1-4b0c-b6b7-8dcb613648fd","Type":"ContainerStarted","Data":"2e248d217af04e0c903894617f3c1ce482d504bf21791be1a0806310ef542cb1"} Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.280764 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="cinder-scheduler" containerID="cri-o://f1820ccfa3e99d47a5ad5a438deabe316bf0a5a2394bb8a85a7e9e190555ddcf" gracePeriod=30 Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.281563 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="probe" containerID="cri-o://fed207cdc0655eed97533254d7521b74e91566b96ddaf69d9876ca818914488b" gracePeriod=30 Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.281780 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gwv89"] Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.303688 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gwv89"] Dec 03 13:02:17 crc kubenswrapper[4990]: I1203 13:02:17.309731 4990 scope.go:117] "RemoveContainer" containerID="273f0647e3e58fbdf3b662556ec4eb22f02fe0f172f8b70e73e29e03ca6b20d4" Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.284887 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" path="/var/lib/kubelet/pods/92763846-ddb6-4b6c-9ebb-8cc0606a4dca/volumes" Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.307888 4990 generic.go:334] "Generic (PLEG): container finished" podID="84d7c974-5413-4a19-bb32-a0baabde0602" containerID="fed207cdc0655eed97533254d7521b74e91566b96ddaf69d9876ca818914488b" exitCode=0 Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.308189 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84d7c974-5413-4a19-bb32-a0baabde0602","Type":"ContainerDied","Data":"fed207cdc0655eed97533254d7521b74e91566b96ddaf69d9876ca818914488b"} Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.319265 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1dd859cc-c778-4e79-8e61-ae6e754e21d7","Type":"ContainerStarted","Data":"0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409"} Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.321819 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65cfc74cf4-n77zl" event={"ID":"630b8058-02f8-4c7e-aeb7-6cca035356ed","Type":"ContainerStarted","Data":"bb3f6561e569f49ce914011f50faae716dc294a3718da2a280b5b41780a35426"} Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.321867 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65cfc74cf4-n77zl" event={"ID":"630b8058-02f8-4c7e-aeb7-6cca035356ed","Type":"ContainerStarted","Data":"0e8c1a6336fdff5125d28eadc45685c9c8e1b9edb6888d482d1e391d468ef3ec"} Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.321914 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.321959 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.325118 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37eba15b-49c1-4b0c-b6b7-8dcb613648fd","Type":"ContainerStarted","Data":"fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827"} Dec 03 13:02:18 crc kubenswrapper[4990]: I1203 13:02:18.350170 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65cfc74cf4-n77zl" podStartSLOduration=3.350147238 podStartE2EDuration="3.350147238s" podCreationTimestamp="2025-12-03 13:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:18.34720922 +0000 UTC m=+1486.489120459" watchObservedRunningTime="2025-12-03 13:02:18.350147238 +0000 UTC m=+1486.492058467" Dec 03 13:02:19 crc kubenswrapper[4990]: I1203 13:02:19.335404 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37eba15b-49c1-4b0c-b6b7-8dcb613648fd","Type":"ContainerStarted","Data":"a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a"} Dec 03 13:02:19 crc kubenswrapper[4990]: I1203 13:02:19.337738 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1dd859cc-c778-4e79-8e61-ae6e754e21d7","Type":"ContainerStarted","Data":"d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8"} Dec 03 13:02:19 crc kubenswrapper[4990]: I1203 13:02:19.354576 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.354557944 podStartE2EDuration="4.354557944s" podCreationTimestamp="2025-12-03 13:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:19.352387847 +0000 UTC m=+1487.494299076" watchObservedRunningTime="2025-12-03 13:02:19.354557944 +0000 UTC m=+1487.496469173" Dec 03 13:02:19 crc kubenswrapper[4990]: I1203 13:02:19.373707 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.373687339 podStartE2EDuration="4.373687339s" podCreationTimestamp="2025-12-03 13:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:19.371812209 +0000 UTC m=+1487.513723438" watchObservedRunningTime="2025-12-03 13:02:19.373687339 +0000 UTC m=+1487.515598568" Dec 03 13:02:20 crc kubenswrapper[4990]: I1203 13:02:20.356578 4990 generic.go:334] "Generic (PLEG): container finished" podID="84d7c974-5413-4a19-bb32-a0baabde0602" containerID="f1820ccfa3e99d47a5ad5a438deabe316bf0a5a2394bb8a85a7e9e190555ddcf" exitCode=0 Dec 03 13:02:20 crc kubenswrapper[4990]: I1203 13:02:20.356620 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84d7c974-5413-4a19-bb32-a0baabde0602","Type":"ContainerDied","Data":"f1820ccfa3e99d47a5ad5a438deabe316bf0a5a2394bb8a85a7e9e190555ddcf"} Dec 03 13:02:20 crc kubenswrapper[4990]: I1203 13:02:20.521774 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6d44745d6d-9zbl5" Dec 03 13:02:20 crc kubenswrapper[4990]: I1203 13:02:20.921525 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:02:20 crc kubenswrapper[4990]: I1203 13:02:20.926974 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.102961 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-secret-key\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103056 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-tls-certs\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103109 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data\") pod \"84d7c974-5413-4a19-bb32-a0baabde0602\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103179 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlfpq\" (UniqueName: \"kubernetes.io/projected/242a176b-2c0d-4067-84a2-48aff8663986-kube-api-access-nlfpq\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103211 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/242a176b-2c0d-4067-84a2-48aff8663986-logs\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103253 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-combined-ca-bundle\") pod \"84d7c974-5413-4a19-bb32-a0baabde0602\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103283 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-scripts\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103322 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data-custom\") pod \"84d7c974-5413-4a19-bb32-a0baabde0602\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103363 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-config-data\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103402 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8xsq\" (UniqueName: \"kubernetes.io/projected/84d7c974-5413-4a19-bb32-a0baabde0602-kube-api-access-n8xsq\") pod \"84d7c974-5413-4a19-bb32-a0baabde0602\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103494 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84d7c974-5413-4a19-bb32-a0baabde0602-etc-machine-id\") pod \"84d7c974-5413-4a19-bb32-a0baabde0602\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103540 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-combined-ca-bundle\") pod \"242a176b-2c0d-4067-84a2-48aff8663986\" (UID: \"242a176b-2c0d-4067-84a2-48aff8663986\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.103572 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-scripts\") pod \"84d7c974-5413-4a19-bb32-a0baabde0602\" (UID: \"84d7c974-5413-4a19-bb32-a0baabde0602\") " Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.105584 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84d7c974-5413-4a19-bb32-a0baabde0602-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "84d7c974-5413-4a19-bb32-a0baabde0602" (UID: "84d7c974-5413-4a19-bb32-a0baabde0602"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.105955 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/242a176b-2c0d-4067-84a2-48aff8663986-logs" (OuterVolumeSpecName: "logs") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.111409 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d7c974-5413-4a19-bb32-a0baabde0602-kube-api-access-n8xsq" (OuterVolumeSpecName: "kube-api-access-n8xsq") pod "84d7c974-5413-4a19-bb32-a0baabde0602" (UID: "84d7c974-5413-4a19-bb32-a0baabde0602"). InnerVolumeSpecName "kube-api-access-n8xsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.111436 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-scripts" (OuterVolumeSpecName: "scripts") pod "84d7c974-5413-4a19-bb32-a0baabde0602" (UID: "84d7c974-5413-4a19-bb32-a0baabde0602"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.115592 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "84d7c974-5413-4a19-bb32-a0baabde0602" (UID: "84d7c974-5413-4a19-bb32-a0baabde0602"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.115662 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.123004 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/242a176b-2c0d-4067-84a2-48aff8663986-kube-api-access-nlfpq" (OuterVolumeSpecName: "kube-api-access-nlfpq") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "kube-api-access-nlfpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.140182 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-config-data" (OuterVolumeSpecName: "config-data") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.145939 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.153270 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-scripts" (OuterVolumeSpecName: "scripts") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.173136 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84d7c974-5413-4a19-bb32-a0baabde0602" (UID: "84d7c974-5413-4a19-bb32-a0baabde0602"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205559 4990 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84d7c974-5413-4a19-bb32-a0baabde0602-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205592 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205602 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205611 4990 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205619 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlfpq\" (UniqueName: \"kubernetes.io/projected/242a176b-2c0d-4067-84a2-48aff8663986-kube-api-access-nlfpq\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205629 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/242a176b-2c0d-4067-84a2-48aff8663986-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205638 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205648 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205656 4990 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205663 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/242a176b-2c0d-4067-84a2-48aff8663986-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.205671 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8xsq\" (UniqueName: \"kubernetes.io/projected/84d7c974-5413-4a19-bb32-a0baabde0602-kube-api-access-n8xsq\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.206276 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "242a176b-2c0d-4067-84a2-48aff8663986" (UID: "242a176b-2c0d-4067-84a2-48aff8663986"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.249743 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data" (OuterVolumeSpecName: "config-data") pod "84d7c974-5413-4a19-bb32-a0baabde0602" (UID: "84d7c974-5413-4a19-bb32-a0baabde0602"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.306826 4990 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/242a176b-2c0d-4067-84a2-48aff8663986-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.306868 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d7c974-5413-4a19-bb32-a0baabde0602-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.366542 4990 generic.go:334] "Generic (PLEG): container finished" podID="242a176b-2c0d-4067-84a2-48aff8663986" containerID="56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213" exitCode=137 Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.366641 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-679fdf78f6-4xbwz" event={"ID":"242a176b-2c0d-4067-84a2-48aff8663986","Type":"ContainerDied","Data":"56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213"} Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.366669 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-679fdf78f6-4xbwz" event={"ID":"242a176b-2c0d-4067-84a2-48aff8663986","Type":"ContainerDied","Data":"ad9815b85abcd52d18b87a28e6293289f84abed9d357a0ced26bf3a4647432b4"} Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.366685 4990 scope.go:117] "RemoveContainer" containerID="b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.366816 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-679fdf78f6-4xbwz" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.384155 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84d7c974-5413-4a19-bb32-a0baabde0602","Type":"ContainerDied","Data":"9481738f7106830fb9b12dc1da6a0b367f51c875cfbcbb2d5f00b2bbe2957f37"} Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.384252 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.519421 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-679fdf78f6-4xbwz"] Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.555554 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-679fdf78f6-4xbwz"] Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.574574 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.598759 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.618781 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.619152 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon-log" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619169 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon-log" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.619193 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="probe" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619200 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="probe" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.619213 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerName="dnsmasq-dns" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619219 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerName="dnsmasq-dns" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.619236 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerName="init" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619246 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerName="init" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.619262 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619270 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.619289 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="cinder-scheduler" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619294 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="cinder-scheduler" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619494 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="92763846-ddb6-4b6c-9ebb-8cc0606a4dca" containerName="dnsmasq-dns" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619522 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="cinder-scheduler" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619532 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon-log" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619539 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" containerName="probe" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.619551 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="242a176b-2c0d-4067-84a2-48aff8663986" containerName="horizon" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.620633 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.621999 4990 scope.go:117] "RemoveContainer" containerID="56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.623937 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.641858 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.647375 4990 scope.go:117] "RemoveContainer" containerID="b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.647839 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa\": container with ID starting with b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa not found: ID does not exist" containerID="b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.647872 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa"} err="failed to get container status \"b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa\": rpc error: code = NotFound desc = could not find container \"b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa\": container with ID starting with b5ff242bd8b0cc89010aa5f7487db71a6fb236329531a01a5690c5b4cf2e54fa not found: ID does not exist" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.647898 4990 scope.go:117] "RemoveContainer" containerID="56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213" Dec 03 13:02:21 crc kubenswrapper[4990]: E1203 13:02:21.648953 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213\": container with ID starting with 56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213 not found: ID does not exist" containerID="56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.648992 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213"} err="failed to get container status \"56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213\": rpc error: code = NotFound desc = could not find container \"56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213\": container with ID starting with 56003863130ecd39541f0ab02abb784cab32b51ae7f956dd9e301b61f603e213 not found: ID does not exist" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.649017 4990 scope.go:117] "RemoveContainer" containerID="fed207cdc0655eed97533254d7521b74e91566b96ddaf69d9876ca818914488b" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.672171 4990 scope.go:117] "RemoveContainer" containerID="f1820ccfa3e99d47a5ad5a438deabe316bf0a5a2394bb8a85a7e9e190555ddcf" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.719133 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ldpp\" (UniqueName: \"kubernetes.io/projected/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-kube-api-access-2ldpp\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.719285 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-scripts\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.719387 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-config-data\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.719546 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.719835 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.719940 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822092 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822143 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822177 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ldpp\" (UniqueName: \"kubernetes.io/projected/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-kube-api-access-2ldpp\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822204 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-scripts\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822230 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-config-data\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822246 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.822315 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.827422 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.827475 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.827749 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-config-data\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.827792 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-scripts\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.837937 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ldpp\" (UniqueName: \"kubernetes.io/projected/d724111e-b7f1-4efe-a05e-1e5cf2f5a278-kube-api-access-2ldpp\") pod \"cinder-scheduler-0\" (UID: \"d724111e-b7f1-4efe-a05e-1e5cf2f5a278\") " pod="openstack/cinder-scheduler-0" Dec 03 13:02:21 crc kubenswrapper[4990]: I1203 13:02:21.949847 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 13:02:22 crc kubenswrapper[4990]: I1203 13:02:22.309403 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="242a176b-2c0d-4067-84a2-48aff8663986" path="/var/lib/kubelet/pods/242a176b-2c0d-4067-84a2-48aff8663986/volumes" Dec 03 13:02:22 crc kubenswrapper[4990]: I1203 13:02:22.310229 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d7c974-5413-4a19-bb32-a0baabde0602" path="/var/lib/kubelet/pods/84d7c974-5413-4a19-bb32-a0baabde0602/volumes" Dec 03 13:02:22 crc kubenswrapper[4990]: I1203 13:02:22.482674 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 13:02:22 crc kubenswrapper[4990]: I1203 13:02:22.674140 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 13:02:23 crc kubenswrapper[4990]: I1203 13:02:23.411773 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d724111e-b7f1-4efe-a05e-1e5cf2f5a278","Type":"ContainerStarted","Data":"86ffe76e572b0fc6f61a7b16cd93beab2c135aff96bef70deba55c6f745c51a4"} Dec 03 13:02:23 crc kubenswrapper[4990]: I1203 13:02:23.412104 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d724111e-b7f1-4efe-a05e-1e5cf2f5a278","Type":"ContainerStarted","Data":"18925854de78eaaa85690257472b2610f67cadb4e6bafc02ac088cae54869263"} Dec 03 13:02:24 crc kubenswrapper[4990]: I1203 13:02:24.422826 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d724111e-b7f1-4efe-a05e-1e5cf2f5a278","Type":"ContainerStarted","Data":"22360d08ab90112748e497c3f26ea808d410c5877d862e18baed125aa38df009"} Dec 03 13:02:24 crc kubenswrapper[4990]: I1203 13:02:24.450122 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.450106264 podStartE2EDuration="3.450106264s" podCreationTimestamp="2025-12-03 13:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:24.446804948 +0000 UTC m=+1492.588716177" watchObservedRunningTime="2025-12-03 13:02:24.450106264 +0000 UTC m=+1492.592017493" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.495019 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.621221 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-86d5555b99-xbxkx"] Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.622668 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.624630 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.635561 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-86d5555b99-xbxkx"] Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.636506 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.640524 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.646907 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.648401 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.663721 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-grf6g" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.664050 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.664110 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.727518 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.739182 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-644b9ffb4d-v4tbk" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.804998 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-public-tls-certs\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805069 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmpzj\" (UniqueName: \"kubernetes.io/projected/62fcb6e6-b158-4a76-8570-3563346e2a77-kube-api-access-dmpzj\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805139 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805173 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fcb6e6-b158-4a76-8570-3563346e2a77-run-httpd\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805204 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pwlp\" (UniqueName: \"kubernetes.io/projected/53fc55c8-2203-45ae-91e5-c377c30a9f20-kube-api-access-6pwlp\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805238 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-combined-ca-bundle\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805280 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-combined-ca-bundle\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805312 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-config-data\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805363 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62fcb6e6-b158-4a76-8570-3563346e2a77-etc-swift\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805394 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fcb6e6-b158-4a76-8570-3563346e2a77-log-httpd\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805418 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config-secret\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.805480 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-internal-tls-certs\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.812742 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.812790 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.861624 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906424 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fcb6e6-b158-4a76-8570-3563346e2a77-run-httpd\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906511 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pwlp\" (UniqueName: \"kubernetes.io/projected/53fc55c8-2203-45ae-91e5-c377c30a9f20-kube-api-access-6pwlp\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906555 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-combined-ca-bundle\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906601 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-combined-ca-bundle\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906638 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-config-data\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906695 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62fcb6e6-b158-4a76-8570-3563346e2a77-etc-swift\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906727 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fcb6e6-b158-4a76-8570-3563346e2a77-log-httpd\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906751 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config-secret\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906794 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-internal-tls-certs\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906831 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-public-tls-certs\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906873 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmpzj\" (UniqueName: \"kubernetes.io/projected/62fcb6e6-b158-4a76-8570-3563346e2a77-kube-api-access-dmpzj\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.906928 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.908180 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.908515 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fcb6e6-b158-4a76-8570-3563346e2a77-log-httpd\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.915166 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-internal-tls-certs\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.917976 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-public-tls-certs\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.922842 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.924131 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fcb6e6-b158-4a76-8570-3563346e2a77-run-httpd\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.934659 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62fcb6e6-b158-4a76-8570-3563346e2a77-etc-swift\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.936442 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-config-data\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.937391 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config-secret\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.938150 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fcb6e6-b158-4a76-8570-3563346e2a77-combined-ca-bundle\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: E1203 13:02:25.943744 4990 projected.go:194] Error preparing data for projected volume kube-api-access-6pwlp for pod openstack/openstackclient: failed to fetch token: pods "openstackclient" not found Dec 03 13:02:25 crc kubenswrapper[4990]: E1203 13:02:25.943849 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53fc55c8-2203-45ae-91e5-c377c30a9f20-kube-api-access-6pwlp podName:53fc55c8-2203-45ae-91e5-c377c30a9f20 nodeName:}" failed. No retries permitted until 2025-12-03 13:02:26.443794233 +0000 UTC m=+1494.585705462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6pwlp" (UniqueName: "kubernetes.io/projected/53fc55c8-2203-45ae-91e5-c377c30a9f20-kube-api-access-6pwlp") pod "openstackclient" (UID: "53fc55c8-2203-45ae-91e5-c377c30a9f20") : failed to fetch token: pods "openstackclient" not found Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.944184 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmpzj\" (UniqueName: \"kubernetes.io/projected/62fcb6e6-b158-4a76-8570-3563346e2a77-kube-api-access-dmpzj\") pod \"swift-proxy-86d5555b99-xbxkx\" (UID: \"62fcb6e6-b158-4a76-8570-3563346e2a77\") " pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.947547 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.948821 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.950018 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.958547 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:25 crc kubenswrapper[4990]: E1203 13:02:25.959274 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-6pwlp], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="53fc55c8-2203-45ae-91e5-c377c30a9f20" Dec 03 13:02:25 crc kubenswrapper[4990]: I1203 13:02:25.964639 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.013110 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-combined-ca-bundle\") pod \"openstackclient\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.043346 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.044855 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.051556 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.062279 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.075930 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.218932 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fdd0021-daac-4e15-b565-25cc720ed808-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.219288 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5gmc\" (UniqueName: \"kubernetes.io/projected/9fdd0021-daac-4e15-b565-25cc720ed808-kube-api-access-l5gmc\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.219327 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdd0021-daac-4e15-b565-25cc720ed808-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.219396 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fdd0021-daac-4e15-b565-25cc720ed808-openstack-config\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.322003 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fdd0021-daac-4e15-b565-25cc720ed808-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.322114 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5gmc\" (UniqueName: \"kubernetes.io/projected/9fdd0021-daac-4e15-b565-25cc720ed808-kube-api-access-l5gmc\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.322160 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdd0021-daac-4e15-b565-25cc720ed808-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.322199 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fdd0021-daac-4e15-b565-25cc720ed808-openstack-config\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.334998 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fdd0021-daac-4e15-b565-25cc720ed808-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.352854 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fdd0021-daac-4e15-b565-25cc720ed808-openstack-config\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.384300 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5gmc\" (UniqueName: \"kubernetes.io/projected/9fdd0021-daac-4e15-b565-25cc720ed808-kube-api-access-l5gmc\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.385076 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fdd0021-daac-4e15-b565-25cc720ed808-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fdd0021-daac-4e15-b565-25cc720ed808\") " pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.435857 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.462639 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.463256 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.463275 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.463284 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.474315 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.494125 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.503382 4990 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="53fc55c8-2203-45ae-91e5-c377c30a9f20" podUID="9fdd0021-daac-4e15-b565-25cc720ed808" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.533193 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config\") pod \"53fc55c8-2203-45ae-91e5-c377c30a9f20\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.533368 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-combined-ca-bundle\") pod \"53fc55c8-2203-45ae-91e5-c377c30a9f20\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.533550 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config-secret\") pod \"53fc55c8-2203-45ae-91e5-c377c30a9f20\" (UID: \"53fc55c8-2203-45ae-91e5-c377c30a9f20\") " Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.534153 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pwlp\" (UniqueName: \"kubernetes.io/projected/53fc55c8-2203-45ae-91e5-c377c30a9f20-kube-api-access-6pwlp\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.535133 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "53fc55c8-2203-45ae-91e5-c377c30a9f20" (UID: "53fc55c8-2203-45ae-91e5-c377c30a9f20"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.547472 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53fc55c8-2203-45ae-91e5-c377c30a9f20" (UID: "53fc55c8-2203-45ae-91e5-c377c30a9f20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.557618 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "53fc55c8-2203-45ae-91e5-c377c30a9f20" (UID: "53fc55c8-2203-45ae-91e5-c377c30a9f20"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.594703 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-86d5555b99-xbxkx"] Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.637071 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.637119 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.637135 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/53fc55c8-2203-45ae-91e5-c377c30a9f20-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:26 crc kubenswrapper[4990]: I1203 13:02:26.951155 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 13:02:27 crc kubenswrapper[4990]: I1203 13:02:27.037606 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 13:02:27 crc kubenswrapper[4990]: I1203 13:02:27.516779 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86d5555b99-xbxkx" event={"ID":"62fcb6e6-b158-4a76-8570-3563346e2a77","Type":"ContainerStarted","Data":"16f594befe11e87be44714a80425afcfd86429eecc68057dd515960c66c8dc25"} Dec 03 13:02:27 crc kubenswrapper[4990]: I1203 13:02:27.517927 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86d5555b99-xbxkx" event={"ID":"62fcb6e6-b158-4a76-8570-3563346e2a77","Type":"ContainerStarted","Data":"3b7f41798bc8689b074630e62c06e93bcc3340ddfe204462a9914b4f104c6b4d"} Dec 03 13:02:27 crc kubenswrapper[4990]: I1203 13:02:27.520708 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9fdd0021-daac-4e15-b565-25cc720ed808","Type":"ContainerStarted","Data":"5e7fb910056a32ac95ea99ff5759da742521d2bc77461f7e535331a4fcb3ec3a"} Dec 03 13:02:27 crc kubenswrapper[4990]: I1203 13:02:27.521060 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 13:02:27 crc kubenswrapper[4990]: I1203 13:02:27.536889 4990 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="53fc55c8-2203-45ae-91e5-c377c30a9f20" podUID="9fdd0021-daac-4e15-b565-25cc720ed808" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.025980 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.030706 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="sg-core" containerID="cri-o://095b85a5e82e9073dd19929e78528edb07f7f9ec551f9e8fd63aaf3900f8882c" gracePeriod=30 Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.030842 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="proxy-httpd" containerID="cri-o://8fce1a54a5833c0dbfce830cda5f4331e0adef18050661c7a4a53eb939e277d6" gracePeriod=30 Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.030909 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-notification-agent" containerID="cri-o://34c01a8173b3088b912702fa17d492d878a741c491fe82c943b7e9d9de79b170" gracePeriod=30 Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.030964 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-central-agent" containerID="cri-o://46657b8f8667997cdee8cdcf0a660fb31ecb761946ae3a86fec6783a83441751" gracePeriod=30 Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.037410 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.288608 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53fc55c8-2203-45ae-91e5-c377c30a9f20" path="/var/lib/kubelet/pods/53fc55c8-2203-45ae-91e5-c377c30a9f20/volumes" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.556113 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86d5555b99-xbxkx" event={"ID":"62fcb6e6-b158-4a76-8570-3563346e2a77","Type":"ContainerStarted","Data":"52e6c0ab0f8a9f41ae96fbe3e9fcb3bd8a4bd18f3ec42ab9172b969da8736cd7"} Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.556497 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.556515 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.586945 4990 generic.go:334] "Generic (PLEG): container finished" podID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerID="8fce1a54a5833c0dbfce830cda5f4331e0adef18050661c7a4a53eb939e277d6" exitCode=0 Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.586977 4990 generic.go:334] "Generic (PLEG): container finished" podID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerID="095b85a5e82e9073dd19929e78528edb07f7f9ec551f9e8fd63aaf3900f8882c" exitCode=2 Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.587040 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.587048 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.587587 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerDied","Data":"8fce1a54a5833c0dbfce830cda5f4331e0adef18050661c7a4a53eb939e277d6"} Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.587623 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerDied","Data":"095b85a5e82e9073dd19929e78528edb07f7f9ec551f9e8fd63aaf3900f8882c"} Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.587668 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.587677 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.588772 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-86d5555b99-xbxkx" podStartSLOduration=3.588763322 podStartE2EDuration="3.588763322s" podCreationTimestamp="2025-12-03 13:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:28.588267979 +0000 UTC m=+1496.730179208" watchObservedRunningTime="2025-12-03 13:02:28.588763322 +0000 UTC m=+1496.730674551" Dec 03 13:02:28 crc kubenswrapper[4990]: I1203 13:02:28.860053 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.273913 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65cfc74cf4-n77zl" Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.353912 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c844ff978-69nft"] Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.354185 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c844ff978-69nft" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api-log" containerID="cri-o://7964d3660976a869710c314952ac1c78a6425a238a2a27a39de137184a7ee7c7" gracePeriod=30 Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.354606 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c844ff978-69nft" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api" containerID="cri-o://0318580cd31a503552c885a434e1c45bd62e99f1ab0c92aab327d47761ca9faa" gracePeriod=30 Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.603275 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.148:3000/\": dial tcp 10.217.0.148:3000: connect: connection refused" Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.627260 4990 generic.go:334] "Generic (PLEG): container finished" podID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerID="7964d3660976a869710c314952ac1c78a6425a238a2a27a39de137184a7ee7c7" exitCode=143 Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.627359 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c844ff978-69nft" event={"ID":"c6a6de82-b949-4b60-b844-1f7c872fe61f","Type":"ContainerDied","Data":"7964d3660976a869710c314952ac1c78a6425a238a2a27a39de137184a7ee7c7"} Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.639444 4990 generic.go:334] "Generic (PLEG): container finished" podID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerID="46657b8f8667997cdee8cdcf0a660fb31ecb761946ae3a86fec6783a83441751" exitCode=0 Dec 03 13:02:29 crc kubenswrapper[4990]: I1203 13:02:29.640509 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerDied","Data":"46657b8f8667997cdee8cdcf0a660fb31ecb761946ae3a86fec6783a83441751"} Dec 03 13:02:30 crc kubenswrapper[4990]: I1203 13:02:30.206335 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:02:30 crc kubenswrapper[4990]: I1203 13:02:30.206493 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:02:30 crc kubenswrapper[4990]: I1203 13:02:30.255198 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:02:30 crc kubenswrapper[4990]: I1203 13:02:30.259091 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:30 crc kubenswrapper[4990]: I1203 13:02:30.259196 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:02:30 crc kubenswrapper[4990]: I1203 13:02:30.262244 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:31 crc kubenswrapper[4990]: I1203 13:02:31.879580 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:32 crc kubenswrapper[4990]: I1203 13:02:32.356429 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 13:02:32 crc kubenswrapper[4990]: I1203 13:02:32.562821 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c844ff978-69nft" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:57160->10.217.0.152:9311: read: connection reset by peer" Dec 03 13:02:32 crc kubenswrapper[4990]: I1203 13:02:32.563578 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c844ff978-69nft" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:57176->10.217.0.152:9311: read: connection reset by peer" Dec 03 13:02:33 crc kubenswrapper[4990]: I1203 13:02:33.286809 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:02:33 crc kubenswrapper[4990]: I1203 13:02:33.286935 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:02:34 crc kubenswrapper[4990]: I1203 13:02:34.711840 4990 generic.go:334] "Generic (PLEG): container finished" podID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerID="34c01a8173b3088b912702fa17d492d878a741c491fe82c943b7e9d9de79b170" exitCode=0 Dec 03 13:02:34 crc kubenswrapper[4990]: I1203 13:02:34.712763 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerDied","Data":"34c01a8173b3088b912702fa17d492d878a741c491fe82c943b7e9d9de79b170"} Dec 03 13:02:34 crc kubenswrapper[4990]: I1203 13:02:34.718319 4990 generic.go:334] "Generic (PLEG): container finished" podID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerID="0318580cd31a503552c885a434e1c45bd62e99f1ab0c92aab327d47761ca9faa" exitCode=0 Dec 03 13:02:34 crc kubenswrapper[4990]: I1203 13:02:34.718531 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c844ff978-69nft" event={"ID":"c6a6de82-b949-4b60-b844-1f7c872fe61f","Type":"ContainerDied","Data":"0318580cd31a503552c885a434e1c45bd62e99f1ab0c92aab327d47761ca9faa"} Dec 03 13:02:35 crc kubenswrapper[4990]: I1203 13:02:35.962473 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:35 crc kubenswrapper[4990]: I1203 13:02:35.963122 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-86d5555b99-xbxkx" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.732741 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.739786 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897681 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-combined-ca-bundle\") pod \"c6a6de82-b949-4b60-b844-1f7c872fe61f\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897723 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-config-data\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897744 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-combined-ca-bundle\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897766 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a6de82-b949-4b60-b844-1f7c872fe61f-logs\") pod \"c6a6de82-b949-4b60-b844-1f7c872fe61f\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897806 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-run-httpd\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897874 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvz86\" (UniqueName: \"kubernetes.io/projected/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-kube-api-access-vvz86\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897893 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-log-httpd\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897916 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rghk\" (UniqueName: \"kubernetes.io/projected/c6a6de82-b949-4b60-b844-1f7c872fe61f-kube-api-access-4rghk\") pod \"c6a6de82-b949-4b60-b844-1f7c872fe61f\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897942 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data\") pod \"c6a6de82-b949-4b60-b844-1f7c872fe61f\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.897973 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data-custom\") pod \"c6a6de82-b949-4b60-b844-1f7c872fe61f\" (UID: \"c6a6de82-b949-4b60-b844-1f7c872fe61f\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.898021 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-sg-core-conf-yaml\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.898052 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-scripts\") pod \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\" (UID: \"8cc2eddc-9da2-4549-bd1d-653a7656ff4a\") " Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.920864 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.925743 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a6de82-b949-4b60-b844-1f7c872fe61f-logs" (OuterVolumeSpecName: "logs") pod "c6a6de82-b949-4b60-b844-1f7c872fe61f" (UID: "c6a6de82-b949-4b60-b844-1f7c872fe61f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.925909 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.943677 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c6a6de82-b949-4b60-b844-1f7c872fe61f" (UID: "c6a6de82-b949-4b60-b844-1f7c872fe61f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.943779 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-scripts" (OuterVolumeSpecName: "scripts") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.948902 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6a6de82-b949-4b60-b844-1f7c872fe61f-kube-api-access-4rghk" (OuterVolumeSpecName: "kube-api-access-4rghk") pod "c6a6de82-b949-4b60-b844-1f7c872fe61f" (UID: "c6a6de82-b949-4b60-b844-1f7c872fe61f"). InnerVolumeSpecName "kube-api-access-4rghk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:40 crc kubenswrapper[4990]: I1203 13:02:40.953633 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-kube-api-access-vvz86" (OuterVolumeSpecName: "kube-api-access-vvz86") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "kube-api-access-vvz86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001662 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001703 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a6de82-b949-4b60-b844-1f7c872fe61f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001724 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001744 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvz86\" (UniqueName: \"kubernetes.io/projected/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-kube-api-access-vvz86\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001757 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001768 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rghk\" (UniqueName: \"kubernetes.io/projected/c6a6de82-b949-4b60-b844-1f7c872fe61f-kube-api-access-4rghk\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.001778 4990 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.005531 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c844ff978-69nft" event={"ID":"c6a6de82-b949-4b60-b844-1f7c872fe61f","Type":"ContainerDied","Data":"96c036f51441633a87eefc277fc2aed5a65ba49ea0125b21ad0b98943f005738"} Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.005579 4990 scope.go:117] "RemoveContainer" containerID="0318580cd31a503552c885a434e1c45bd62e99f1ab0c92aab327d47761ca9faa" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.005714 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c844ff978-69nft" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.027618 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8cc2eddc-9da2-4549-bd1d-653a7656ff4a","Type":"ContainerDied","Data":"72a639c71b273fcf0ddb5128fd6e228677f5d5633b4d7e6bc0cfaf05d8bc6821"} Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.027735 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.033825 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.044621 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6a6de82-b949-4b60-b844-1f7c872fe61f" (UID: "c6a6de82-b949-4b60-b844-1f7c872fe61f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.064433 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data" (OuterVolumeSpecName: "config-data") pod "c6a6de82-b949-4b60-b844-1f7c872fe61f" (UID: "c6a6de82-b949-4b60-b844-1f7c872fe61f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.066653 4990 scope.go:117] "RemoveContainer" containerID="7964d3660976a869710c314952ac1c78a6425a238a2a27a39de137184a7ee7c7" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.102590 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.103918 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.103951 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.103962 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a6de82-b949-4b60-b844-1f7c872fe61f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.103974 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.117648 4990 scope.go:117] "RemoveContainer" containerID="8fce1a54a5833c0dbfce830cda5f4331e0adef18050661c7a4a53eb939e277d6" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.143320 4990 scope.go:117] "RemoveContainer" containerID="095b85a5e82e9073dd19929e78528edb07f7f9ec551f9e8fd63aaf3900f8882c" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.161925 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-config-data" (OuterVolumeSpecName: "config-data") pod "8cc2eddc-9da2-4549-bd1d-653a7656ff4a" (UID: "8cc2eddc-9da2-4549-bd1d-653a7656ff4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.175399 4990 scope.go:117] "RemoveContainer" containerID="34c01a8173b3088b912702fa17d492d878a741c491fe82c943b7e9d9de79b170" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.195281 4990 scope.go:117] "RemoveContainer" containerID="46657b8f8667997cdee8cdcf0a660fb31ecb761946ae3a86fec6783a83441751" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.205292 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cc2eddc-9da2-4549-bd1d-653a7656ff4a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.340899 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c844ff978-69nft"] Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.340983 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c844ff978-69nft" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.341096 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c844ff978-69nft" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.346325 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c844ff978-69nft"] Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.363729 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.372395 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.388800 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:41 crc kubenswrapper[4990]: E1203 13:02:41.389242 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="proxy-httpd" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389266 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="proxy-httpd" Dec 03 13:02:41 crc kubenswrapper[4990]: E1203 13:02:41.389284 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api-log" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389292 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api-log" Dec 03 13:02:41 crc kubenswrapper[4990]: E1203 13:02:41.389304 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-notification-agent" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389313 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-notification-agent" Dec 03 13:02:41 crc kubenswrapper[4990]: E1203 13:02:41.389328 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389336 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api" Dec 03 13:02:41 crc kubenswrapper[4990]: E1203 13:02:41.389357 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-central-agent" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389365 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-central-agent" Dec 03 13:02:41 crc kubenswrapper[4990]: E1203 13:02:41.389377 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="sg-core" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389385 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="sg-core" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389598 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389623 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="proxy-httpd" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389634 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="sg-core" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389645 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" containerName="barbican-api-log" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389656 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-notification-agent" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.389669 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" containerName="ceilometer-central-agent" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.396877 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.407957 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.413743 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.413977 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.510947 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-run-httpd\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.511039 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9smsl\" (UniqueName: \"kubernetes.io/projected/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-kube-api-access-9smsl\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.511330 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-config-data\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.511388 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-log-httpd\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.511410 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.511430 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.511503 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-scripts\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613512 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-log-httpd\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613568 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613594 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613630 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-scripts\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613700 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-run-httpd\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613737 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9smsl\" (UniqueName: \"kubernetes.io/projected/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-kube-api-access-9smsl\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.613769 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-config-data\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.614149 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-log-httpd\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.614774 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-run-httpd\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.618259 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.618605 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-scripts\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.618834 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-config-data\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.623613 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.637139 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9smsl\" (UniqueName: \"kubernetes.io/projected/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-kube-api-access-9smsl\") pod \"ceilometer-0\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " pod="openstack/ceilometer-0" Dec 03 13:02:41 crc kubenswrapper[4990]: I1203 13:02:41.729698 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:42 crc kubenswrapper[4990]: I1203 13:02:42.039524 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9fdd0021-daac-4e15-b565-25cc720ed808","Type":"ContainerStarted","Data":"14a99227ec7cb42e8dbe4be98bfd49228859e4bfe630476957d431d444a01fd4"} Dec 03 13:02:42 crc kubenswrapper[4990]: I1203 13:02:42.066318 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.319981905 podStartE2EDuration="17.066296257s" podCreationTimestamp="2025-12-03 13:02:25 +0000 UTC" firstStartedPulling="2025-12-03 13:02:27.078312722 +0000 UTC m=+1495.220223951" lastFinishedPulling="2025-12-03 13:02:40.824627074 +0000 UTC m=+1508.966538303" observedRunningTime="2025-12-03 13:02:42.058528013 +0000 UTC m=+1510.200439252" watchObservedRunningTime="2025-12-03 13:02:42.066296257 +0000 UTC m=+1510.208207486" Dec 03 13:02:42 crc kubenswrapper[4990]: I1203 13:02:42.223389 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:42 crc kubenswrapper[4990]: I1203 13:02:42.277872 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc2eddc-9da2-4549-bd1d-653a7656ff4a" path="/var/lib/kubelet/pods/8cc2eddc-9da2-4549-bd1d-653a7656ff4a/volumes" Dec 03 13:02:42 crc kubenswrapper[4990]: I1203 13:02:42.278867 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6a6de82-b949-4b60-b844-1f7c872fe61f" path="/var/lib/kubelet/pods/c6a6de82-b949-4b60-b844-1f7c872fe61f/volumes" Dec 03 13:02:43 crc kubenswrapper[4990]: I1203 13:02:43.055009 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerStarted","Data":"9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016"} Dec 03 13:02:43 crc kubenswrapper[4990]: I1203 13:02:43.055303 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerStarted","Data":"32beec4ebbb056febeb5139b9c9e7b3a991edd2c2267322ea4e029f904a570d7"} Dec 03 13:02:43 crc kubenswrapper[4990]: I1203 13:02:43.114424 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69b4f6458f-phdfw" Dec 03 13:02:43 crc kubenswrapper[4990]: I1203 13:02:43.184510 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d4d966994-44d9g"] Dec 03 13:02:43 crc kubenswrapper[4990]: I1203 13:02:43.184836 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d4d966994-44d9g" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-api" containerID="cri-o://311f3a153e8a397f7a5dbb6a13f6df38c9cfab799395632057389212ab43d43a" gracePeriod=30 Dec 03 13:02:43 crc kubenswrapper[4990]: I1203 13:02:43.184972 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d4d966994-44d9g" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-httpd" containerID="cri-o://cb6274e5fb2c79285399004898ea1e83c59b93160a26ddfe2f391cfd7b839ccc" gracePeriod=30 Dec 03 13:02:44 crc kubenswrapper[4990]: I1203 13:02:44.076653 4990 generic.go:334] "Generic (PLEG): container finished" podID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerID="cb6274e5fb2c79285399004898ea1e83c59b93160a26ddfe2f391cfd7b839ccc" exitCode=0 Dec 03 13:02:44 crc kubenswrapper[4990]: I1203 13:02:44.076716 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d4d966994-44d9g" event={"ID":"cab24acb-b936-4302-88ba-1ccf6f29d81e","Type":"ContainerDied","Data":"cb6274e5fb2c79285399004898ea1e83c59b93160a26ddfe2f391cfd7b839ccc"} Dec 03 13:02:44 crc kubenswrapper[4990]: I1203 13:02:44.080256 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerStarted","Data":"95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30"} Dec 03 13:02:44 crc kubenswrapper[4990]: I1203 13:02:44.330806 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:45 crc kubenswrapper[4990]: I1203 13:02:45.097127 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerStarted","Data":"c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63"} Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.110331 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerStarted","Data":"85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623"} Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.110553 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="proxy-httpd" containerID="cri-o://85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623" gracePeriod=30 Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.110604 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-notification-agent" containerID="cri-o://95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30" gracePeriod=30 Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.110772 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.111131 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-central-agent" containerID="cri-o://9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016" gracePeriod=30 Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.110662 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="sg-core" containerID="cri-o://c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63" gracePeriod=30 Dec 03 13:02:46 crc kubenswrapper[4990]: I1203 13:02:46.146154 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8769619149999999 podStartE2EDuration="5.146135423s" podCreationTimestamp="2025-12-03 13:02:41 +0000 UTC" firstStartedPulling="2025-12-03 13:02:42.230626351 +0000 UTC m=+1510.372537580" lastFinishedPulling="2025-12-03 13:02:45.499799869 +0000 UTC m=+1513.641711088" observedRunningTime="2025-12-03 13:02:46.141057369 +0000 UTC m=+1514.282968598" watchObservedRunningTime="2025-12-03 13:02:46.146135423 +0000 UTC m=+1514.288046662" Dec 03 13:02:47 crc kubenswrapper[4990]: I1203 13:02:47.142387 4990 generic.go:334] "Generic (PLEG): container finished" podID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerID="85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623" exitCode=0 Dec 03 13:02:47 crc kubenswrapper[4990]: I1203 13:02:47.142722 4990 generic.go:334] "Generic (PLEG): container finished" podID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerID="c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63" exitCode=2 Dec 03 13:02:47 crc kubenswrapper[4990]: I1203 13:02:47.142751 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerDied","Data":"85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623"} Dec 03 13:02:47 crc kubenswrapper[4990]: I1203 13:02:47.142780 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerDied","Data":"c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63"} Dec 03 13:02:50 crc kubenswrapper[4990]: I1203 13:02:50.179527 4990 generic.go:334] "Generic (PLEG): container finished" podID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerID="95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30" exitCode=0 Dec 03 13:02:50 crc kubenswrapper[4990]: I1203 13:02:50.179605 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerDied","Data":"95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30"} Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.191408 4990 generic.go:334] "Generic (PLEG): container finished" podID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerID="311f3a153e8a397f7a5dbb6a13f6df38c9cfab799395632057389212ab43d43a" exitCode=0 Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.191729 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d4d966994-44d9g" event={"ID":"cab24acb-b936-4302-88ba-1ccf6f29d81e","Type":"ContainerDied","Data":"311f3a153e8a397f7a5dbb6a13f6df38c9cfab799395632057389212ab43d43a"} Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.191756 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d4d966994-44d9g" event={"ID":"cab24acb-b936-4302-88ba-1ccf6f29d81e","Type":"ContainerDied","Data":"cb35b7994990d10c29427bdba14365642a6dfad6b885051f0532d8a05fbc227d"} Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.191766 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb35b7994990d10c29427bdba14365642a6dfad6b885051f0532d8a05fbc227d" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.290998 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.411237 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-ovndb-tls-certs\") pod \"cab24acb-b936-4302-88ba-1ccf6f29d81e\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.411359 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67gzf\" (UniqueName: \"kubernetes.io/projected/cab24acb-b936-4302-88ba-1ccf6f29d81e-kube-api-access-67gzf\") pod \"cab24acb-b936-4302-88ba-1ccf6f29d81e\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.411497 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-config\") pod \"cab24acb-b936-4302-88ba-1ccf6f29d81e\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.411743 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-combined-ca-bundle\") pod \"cab24acb-b936-4302-88ba-1ccf6f29d81e\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.411811 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-httpd-config\") pod \"cab24acb-b936-4302-88ba-1ccf6f29d81e\" (UID: \"cab24acb-b936-4302-88ba-1ccf6f29d81e\") " Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.416827 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "cab24acb-b936-4302-88ba-1ccf6f29d81e" (UID: "cab24acb-b936-4302-88ba-1ccf6f29d81e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.426351 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab24acb-b936-4302-88ba-1ccf6f29d81e-kube-api-access-67gzf" (OuterVolumeSpecName: "kube-api-access-67gzf") pod "cab24acb-b936-4302-88ba-1ccf6f29d81e" (UID: "cab24acb-b936-4302-88ba-1ccf6f29d81e"). InnerVolumeSpecName "kube-api-access-67gzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.469046 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cab24acb-b936-4302-88ba-1ccf6f29d81e" (UID: "cab24acb-b936-4302-88ba-1ccf6f29d81e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.475363 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-config" (OuterVolumeSpecName: "config") pod "cab24acb-b936-4302-88ba-1ccf6f29d81e" (UID: "cab24acb-b936-4302-88ba-1ccf6f29d81e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.498535 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "cab24acb-b936-4302-88ba-1ccf6f29d81e" (UID: "cab24acb-b936-4302-88ba-1ccf6f29d81e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.513784 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.513811 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.513821 4990 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.513831 4990 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab24acb-b936-4302-88ba-1ccf6f29d81e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.513839 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67gzf\" (UniqueName: \"kubernetes.io/projected/cab24acb-b936-4302-88ba-1ccf6f29d81e-kube-api-access-67gzf\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.742998 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.743234 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-log" containerID="cri-o://fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827" gracePeriod=30 Dec 03 13:02:51 crc kubenswrapper[4990]: I1203 13:02:51.743300 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-httpd" containerID="cri-o://a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a" gracePeriod=30 Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.202930 4990 generic.go:334] "Generic (PLEG): container finished" podID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerID="fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827" exitCode=143 Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.203821 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d4d966994-44d9g" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.212327 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37eba15b-49c1-4b0c-b6b7-8dcb613648fd","Type":"ContainerDied","Data":"fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827"} Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.235056 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d4d966994-44d9g"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.247106 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7d4d966994-44d9g"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.275542 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" path="/var/lib/kubelet/pods/cab24acb-b936-4302-88ba-1ccf6f29d81e/volumes" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.577613 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-54jvk"] Dec 03 13:02:52 crc kubenswrapper[4990]: E1203 13:02:52.578076 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-api" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.578101 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-api" Dec 03 13:02:52 crc kubenswrapper[4990]: E1203 13:02:52.578124 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-httpd" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.578132 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-httpd" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.578346 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-httpd" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.578367 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab24acb-b936-4302-88ba-1ccf6f29d81e" containerName="neutron-api" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.579141 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.591833 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-54jvk"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.686890 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-znfpn"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.688184 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.701861 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3383-account-create-update-6ggwh"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.704163 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.707382 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.720054 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-znfpn"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.737936 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3383-account-create-update-6ggwh"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.739732 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67pbs\" (UniqueName: \"kubernetes.io/projected/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-kube-api-access-67pbs\") pod \"nova-api-db-create-54jvk\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.739915 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-operator-scripts\") pod \"nova-api-db-create-54jvk\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.785374 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-7mtkf"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.788887 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.792969 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7mtkf"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.841708 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67pbs\" (UniqueName: \"kubernetes.io/projected/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-kube-api-access-67pbs\") pod \"nova-api-db-create-54jvk\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.841763 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khfkh\" (UniqueName: \"kubernetes.io/projected/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-kube-api-access-khfkh\") pod \"nova-api-3383-account-create-update-6ggwh\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.841788 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d4ddf00-b53e-4d9c-bfaa-676226812d86-operator-scripts\") pod \"nova-cell0-db-create-znfpn\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.841814 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-operator-scripts\") pod \"nova-api-3383-account-create-update-6ggwh\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.841905 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-operator-scripts\") pod \"nova-api-db-create-54jvk\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.841957 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bsrq\" (UniqueName: \"kubernetes.io/projected/8d4ddf00-b53e-4d9c-bfaa-676226812d86-kube-api-access-4bsrq\") pod \"nova-cell0-db-create-znfpn\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.843170 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-operator-scripts\") pod \"nova-api-db-create-54jvk\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.863820 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67pbs\" (UniqueName: \"kubernetes.io/projected/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-kube-api-access-67pbs\") pod \"nova-api-db-create-54jvk\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.880324 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6d20-account-create-update-929g9"] Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.892441 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.894911 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.935108 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.943088 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjh6j\" (UniqueName: \"kubernetes.io/projected/2aceea00-aeff-45f4-83ae-c7656a28be74-kube-api-access-vjh6j\") pod \"nova-cell1-db-create-7mtkf\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.943257 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bsrq\" (UniqueName: \"kubernetes.io/projected/8d4ddf00-b53e-4d9c-bfaa-676226812d86-kube-api-access-4bsrq\") pod \"nova-cell0-db-create-znfpn\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.943364 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aceea00-aeff-45f4-83ae-c7656a28be74-operator-scripts\") pod \"nova-cell1-db-create-7mtkf\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.943413 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khfkh\" (UniqueName: \"kubernetes.io/projected/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-kube-api-access-khfkh\") pod \"nova-api-3383-account-create-update-6ggwh\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.943445 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d4ddf00-b53e-4d9c-bfaa-676226812d86-operator-scripts\") pod \"nova-cell0-db-create-znfpn\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.943494 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-operator-scripts\") pod \"nova-api-3383-account-create-update-6ggwh\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.944246 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d4ddf00-b53e-4d9c-bfaa-676226812d86-operator-scripts\") pod \"nova-cell0-db-create-znfpn\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.944499 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-operator-scripts\") pod \"nova-api-3383-account-create-update-6ggwh\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.976045 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khfkh\" (UniqueName: \"kubernetes.io/projected/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-kube-api-access-khfkh\") pod \"nova-api-3383-account-create-update-6ggwh\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:52 crc kubenswrapper[4990]: I1203 13:02:52.977504 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bsrq\" (UniqueName: \"kubernetes.io/projected/8d4ddf00-b53e-4d9c-bfaa-676226812d86-kube-api-access-4bsrq\") pod \"nova-cell0-db-create-znfpn\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.011417 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.031533 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.045025 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aceea00-aeff-45f4-83ae-c7656a28be74-operator-scripts\") pod \"nova-cell1-db-create-7mtkf\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.045344 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjh6j\" (UniqueName: \"kubernetes.io/projected/2aceea00-aeff-45f4-83ae-c7656a28be74-kube-api-access-vjh6j\") pod \"nova-cell1-db-create-7mtkf\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.045372 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnxmv\" (UniqueName: \"kubernetes.io/projected/a7a46743-abd7-4cc7-9594-60332bb6aa8d-kube-api-access-wnxmv\") pod \"nova-cell0-6d20-account-create-update-929g9\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.045462 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a46743-abd7-4cc7-9594-60332bb6aa8d-operator-scripts\") pod \"nova-cell0-6d20-account-create-update-929g9\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.045912 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aceea00-aeff-45f4-83ae-c7656a28be74-operator-scripts\") pod \"nova-cell1-db-create-7mtkf\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.087166 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjh6j\" (UniqueName: \"kubernetes.io/projected/2aceea00-aeff-45f4-83ae-c7656a28be74-kube-api-access-vjh6j\") pod \"nova-cell1-db-create-7mtkf\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.090473 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5857-account-create-update-bf22s"] Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.093701 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.096889 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.110443 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.124571 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5857-account-create-update-bf22s"] Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.146536 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnxmv\" (UniqueName: \"kubernetes.io/projected/a7a46743-abd7-4cc7-9594-60332bb6aa8d-kube-api-access-wnxmv\") pod \"nova-cell0-6d20-account-create-update-929g9\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.146616 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a46743-abd7-4cc7-9594-60332bb6aa8d-operator-scripts\") pod \"nova-cell0-6d20-account-create-update-929g9\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.147294 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a46743-abd7-4cc7-9594-60332bb6aa8d-operator-scripts\") pod \"nova-cell0-6d20-account-create-update-929g9\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.166193 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnxmv\" (UniqueName: \"kubernetes.io/projected/a7a46743-abd7-4cc7-9594-60332bb6aa8d-kube-api-access-wnxmv\") pod \"nova-cell0-6d20-account-create-update-929g9\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.214019 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.248133 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmzrv\" (UniqueName: \"kubernetes.io/projected/1b129617-7cab-45cc-9cb0-00fdf619bec2-kube-api-access-rmzrv\") pod \"nova-cell1-5857-account-create-update-bf22s\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.248416 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b129617-7cab-45cc-9cb0-00fdf619bec2-operator-scripts\") pod \"nova-cell1-5857-account-create-update-bf22s\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.349891 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmzrv\" (UniqueName: \"kubernetes.io/projected/1b129617-7cab-45cc-9cb0-00fdf619bec2-kube-api-access-rmzrv\") pod \"nova-cell1-5857-account-create-update-bf22s\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.349950 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b129617-7cab-45cc-9cb0-00fdf619bec2-operator-scripts\") pod \"nova-cell1-5857-account-create-update-bf22s\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.351067 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b129617-7cab-45cc-9cb0-00fdf619bec2-operator-scripts\") pod \"nova-cell1-5857-account-create-update-bf22s\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.365940 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmzrv\" (UniqueName: \"kubernetes.io/projected/1b129617-7cab-45cc-9cb0-00fdf619bec2-kube-api-access-rmzrv\") pod \"nova-cell1-5857-account-create-update-bf22s\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.470859 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.479075 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-54jvk"] Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.609105 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3383-account-create-update-6ggwh"] Dec 03 13:02:53 crc kubenswrapper[4990]: W1203 13:02:53.623637 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4e36ec_4ef9_4773_8ae2_2b0dabf61b3e.slice/crio-6a5dba92e88ef4defb613815cd4b1ad1bf2cda4b05faf4ff48f916b26743acaf WatchSource:0}: Error finding container 6a5dba92e88ef4defb613815cd4b1ad1bf2cda4b05faf4ff48f916b26743acaf: Status 404 returned error can't find the container with id 6a5dba92e88ef4defb613815cd4b1ad1bf2cda4b05faf4ff48f916b26743acaf Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.624657 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7mtkf"] Dec 03 13:02:53 crc kubenswrapper[4990]: W1203 13:02:53.630825 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aceea00_aeff_45f4_83ae_c7656a28be74.slice/crio-abe1dec3c6cee2644327582ef0182de9ec76c421f7909b354acef343f16c050d WatchSource:0}: Error finding container abe1dec3c6cee2644327582ef0182de9ec76c421f7909b354acef343f16c050d: Status 404 returned error can't find the container with id abe1dec3c6cee2644327582ef0182de9ec76c421f7909b354acef343f16c050d Dec 03 13:02:53 crc kubenswrapper[4990]: I1203 13:02:53.910409 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5857-account-create-update-bf22s"] Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.232878 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3383-account-create-update-6ggwh" event={"ID":"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e","Type":"ContainerStarted","Data":"6a5dba92e88ef4defb613815cd4b1ad1bf2cda4b05faf4ff48f916b26743acaf"} Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.234302 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5857-account-create-update-bf22s" event={"ID":"1b129617-7cab-45cc-9cb0-00fdf619bec2","Type":"ContainerStarted","Data":"59447b9c38b66008cf6194e759dabe8ad03d7ca0a7b0e946ce0434a664a9821c"} Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.235813 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7mtkf" event={"ID":"2aceea00-aeff-45f4-83ae-c7656a28be74","Type":"ContainerStarted","Data":"abe1dec3c6cee2644327582ef0182de9ec76c421f7909b354acef343f16c050d"} Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.237242 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-54jvk" event={"ID":"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27","Type":"ContainerStarted","Data":"340f7bfa5ad96ffd7320001276b5aefeb735a38ff585e2cd4c4ddf4c7e19cb0d"} Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.597425 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6d20-account-create-update-929g9"] Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.929023 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.929483 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-log" containerID="cri-o://0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409" gracePeriod=30 Dec 03 13:02:54 crc kubenswrapper[4990]: I1203 13:02:54.930052 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-httpd" containerID="cri-o://d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8" gracePeriod=30 Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.018836 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-znfpn"] Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.216017 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6d20-account-create-update-929g9"] Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.264103 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6d20-account-create-update-929g9" event={"ID":"a7a46743-abd7-4cc7-9594-60332bb6aa8d","Type":"ContainerStarted","Data":"6c0dfcdabb4d76e4f9b9c7f26d470885478919d53903cd4f91455c230a7f1762"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.274319 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7mtkf" event={"ID":"2aceea00-aeff-45f4-83ae-c7656a28be74","Type":"ContainerStarted","Data":"b4bdde7dd51e54840effae84404aa7d83e40ded8774a607e6f79223cd7052989"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.281521 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-znfpn" event={"ID":"8d4ddf00-b53e-4d9c-bfaa-676226812d86","Type":"ContainerStarted","Data":"d56ef90c0b8f397f600412dfb75e9082bb7788b8d5376efd39aa7617eedfde85"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.286678 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-54jvk" event={"ID":"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27","Type":"ContainerStarted","Data":"f5451a9033146b5ec9114fd6279daf4099c6deb267cf5b1ff67747226bdfa42d"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.293257 4990 generic.go:334] "Generic (PLEG): container finished" podID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerID="0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409" exitCode=143 Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.293355 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1dd859cc-c778-4e79-8e61-ae6e754e21d7","Type":"ContainerDied","Data":"0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.294967 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3383-account-create-update-6ggwh" event={"ID":"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e","Type":"ContainerStarted","Data":"0aaeac0d890447ab06b1675bbc0ecd452feed1f9651d741ea7b8a363829e68d3"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.300374 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5857-account-create-update-bf22s" event={"ID":"1b129617-7cab-45cc-9cb0-00fdf619bec2","Type":"ContainerStarted","Data":"fee086a7a2fe21abcad12fe048099b83ad057b3b2f7c3b7157205f2850508c83"} Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.305783 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-7mtkf" podStartSLOduration=3.305764215 podStartE2EDuration="3.305764215s" podCreationTimestamp="2025-12-03 13:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:55.303575097 +0000 UTC m=+1523.445486336" watchObservedRunningTime="2025-12-03 13:02:55.305764215 +0000 UTC m=+1523.447675444" Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.322713 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-54jvk" podStartSLOduration=3.322691651 podStartE2EDuration="3.322691651s" podCreationTimestamp="2025-12-03 13:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:55.320969776 +0000 UTC m=+1523.462881015" watchObservedRunningTime="2025-12-03 13:02:55.322691651 +0000 UTC m=+1523.464602880" Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.338442 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-3383-account-create-update-6ggwh" podStartSLOduration=3.338422456 podStartE2EDuration="3.338422456s" podCreationTimestamp="2025-12-03 13:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:55.332184662 +0000 UTC m=+1523.474095891" watchObservedRunningTime="2025-12-03 13:02:55.338422456 +0000 UTC m=+1523.480333685" Dec 03 13:02:55 crc kubenswrapper[4990]: I1203 13:02:55.351886 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5857-account-create-update-bf22s" podStartSLOduration=2.351868281 podStartE2EDuration="2.351868281s" podCreationTimestamp="2025-12-03 13:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:55.348181073 +0000 UTC m=+1523.490092322" watchObservedRunningTime="2025-12-03 13:02:55.351868281 +0000 UTC m=+1523.493779510" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.082080 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207594 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-httpd-run\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207658 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-public-tls-certs\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207700 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmm48\" (UniqueName: \"kubernetes.io/projected/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-kube-api-access-gmm48\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207831 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-config-data\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207868 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-scripts\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207922 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-logs\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207944 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.207961 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-combined-ca-bundle\") pod \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\" (UID: \"37eba15b-49c1-4b0c-b6b7-8dcb613648fd\") " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.208124 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.208796 4990 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.210125 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-logs" (OuterVolumeSpecName: "logs") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.230128 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-scripts" (OuterVolumeSpecName: "scripts") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.232671 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.246578 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-kube-api-access-gmm48" (OuterVolumeSpecName: "kube-api-access-gmm48") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "kube-api-access-gmm48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.265443 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.269369 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-config-data" (OuterVolumeSpecName: "config-data") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.290067 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "37eba15b-49c1-4b0c-b6b7-8dcb613648fd" (UID: "37eba15b-49c1-4b0c-b6b7-8dcb613648fd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310312 4990 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310343 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmm48\" (UniqueName: \"kubernetes.io/projected/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-kube-api-access-gmm48\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310354 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310362 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310374 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310409 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.310419 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eba15b-49c1-4b0c-b6b7-8dcb613648fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.323661 4990 generic.go:334] "Generic (PLEG): container finished" podID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerID="a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a" exitCode=0 Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.323848 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.346555 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.361280 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-6d20-account-create-update-929g9" podStartSLOduration=4.361257309 podStartE2EDuration="4.361257309s" podCreationTimestamp="2025-12-03 13:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:56.346109769 +0000 UTC m=+1524.488021008" watchObservedRunningTime="2025-12-03 13:02:56.361257309 +0000 UTC m=+1524.503168538" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.367159 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-znfpn" podStartSLOduration=4.367144654 podStartE2EDuration="4.367144654s" podCreationTimestamp="2025-12-03 13:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:56.36359528 +0000 UTC m=+1524.505506509" watchObservedRunningTime="2025-12-03 13:02:56.367144654 +0000 UTC m=+1524.509055873" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.393348 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37eba15b-49c1-4b0c-b6b7-8dcb613648fd","Type":"ContainerDied","Data":"a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a"} Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.393406 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37eba15b-49c1-4b0c-b6b7-8dcb613648fd","Type":"ContainerDied","Data":"2e248d217af04e0c903894617f3c1ce482d504bf21791be1a0806310ef542cb1"} Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.393423 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6d20-account-create-update-929g9" event={"ID":"a7a46743-abd7-4cc7-9594-60332bb6aa8d","Type":"ContainerStarted","Data":"8a62221fe96a9eb71de6e3730b0bc15c912092e90565f1cbfc86aed4918e7823"} Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.393436 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-znfpn" event={"ID":"8d4ddf00-b53e-4d9c-bfaa-676226812d86","Type":"ContainerStarted","Data":"4a51ba2b9557f90c165c3ca6289b1d5342f36d2366e0cad02d656032cf556801"} Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.393474 4990 scope.go:117] "RemoveContainer" containerID="a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.414231 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.423014 4990 scope.go:117] "RemoveContainer" containerID="fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.428599 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.438139 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.450119 4990 scope.go:117] "RemoveContainer" containerID="a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a" Dec 03 13:02:56 crc kubenswrapper[4990]: E1203 13:02:56.450766 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a\": container with ID starting with a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a not found: ID does not exist" containerID="a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.450806 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a"} err="failed to get container status \"a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a\": rpc error: code = NotFound desc = could not find container \"a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a\": container with ID starting with a2e68842f87756fe321dbe42c90a5d6fe30c061b46c4e847061adace0f9b895a not found: ID does not exist" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.450831 4990 scope.go:117] "RemoveContainer" containerID="fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827" Dec 03 13:02:56 crc kubenswrapper[4990]: E1203 13:02:56.452914 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827\": container with ID starting with fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827 not found: ID does not exist" containerID="fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.452952 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827"} err="failed to get container status \"fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827\": rpc error: code = NotFound desc = could not find container \"fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827\": container with ID starting with fd6476b04a55d21bd0e2e650e0fbc2b804d570fff9e786ecb0f3cf9d48021827 not found: ID does not exist" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.454151 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:56 crc kubenswrapper[4990]: E1203 13:02:56.454708 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-log" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.454729 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-log" Dec 03 13:02:56 crc kubenswrapper[4990]: E1203 13:02:56.454764 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-httpd" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.454775 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-httpd" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.455110 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-log" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.455147 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" containerName="glance-httpd" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.456652 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.461539 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.461715 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.469945 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515285 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515346 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-logs\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515367 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515388 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515407 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np2bm\" (UniqueName: \"kubernetes.io/projected/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-kube-api-access-np2bm\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515433 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515496 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.515531 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.617098 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-logs\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.617414 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.617565 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.617677 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np2bm\" (UniqueName: \"kubernetes.io/projected/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-kube-api-access-np2bm\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.617798 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.617937 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.618064 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.618275 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.618884 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.619336 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-logs\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.621239 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.629409 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.629640 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.632497 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.643801 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.644567 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np2bm\" (UniqueName: \"kubernetes.io/projected/fbdb7170-e955-4fd4-bfd5-03ad13363aa9-kube-api-access-np2bm\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.662345 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"fbdb7170-e955-4fd4-bfd5-03ad13363aa9\") " pod="openstack/glance-default-external-api-0" Dec 03 13:02:56 crc kubenswrapper[4990]: I1203 13:02:56.788389 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.330223 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.381835 4990 generic.go:334] "Generic (PLEG): container finished" podID="1b129617-7cab-45cc-9cb0-00fdf619bec2" containerID="fee086a7a2fe21abcad12fe048099b83ad057b3b2f7c3b7157205f2850508c83" exitCode=0 Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.381910 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5857-account-create-update-bf22s" event={"ID":"1b129617-7cab-45cc-9cb0-00fdf619bec2","Type":"ContainerDied","Data":"fee086a7a2fe21abcad12fe048099b83ad057b3b2f7c3b7157205f2850508c83"} Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.388595 4990 generic.go:334] "Generic (PLEG): container finished" podID="2aceea00-aeff-45f4-83ae-c7656a28be74" containerID="b4bdde7dd51e54840effae84404aa7d83e40ded8774a607e6f79223cd7052989" exitCode=0 Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.388677 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7mtkf" event={"ID":"2aceea00-aeff-45f4-83ae-c7656a28be74","Type":"ContainerDied","Data":"b4bdde7dd51e54840effae84404aa7d83e40ded8774a607e6f79223cd7052989"} Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.392388 4990 generic.go:334] "Generic (PLEG): container finished" podID="8d4ddf00-b53e-4d9c-bfaa-676226812d86" containerID="4a51ba2b9557f90c165c3ca6289b1d5342f36d2366e0cad02d656032cf556801" exitCode=0 Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.392606 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-znfpn" event={"ID":"8d4ddf00-b53e-4d9c-bfaa-676226812d86","Type":"ContainerDied","Data":"4a51ba2b9557f90c165c3ca6289b1d5342f36d2366e0cad02d656032cf556801"} Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.394470 4990 generic.go:334] "Generic (PLEG): container finished" podID="1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" containerID="f5451a9033146b5ec9114fd6279daf4099c6deb267cf5b1ff67747226bdfa42d" exitCode=0 Dec 03 13:02:57 crc kubenswrapper[4990]: I1203 13:02:57.394609 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-54jvk" event={"ID":"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27","Type":"ContainerDied","Data":"f5451a9033146b5ec9114fd6279daf4099c6deb267cf5b1ff67747226bdfa42d"} Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.278345 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37eba15b-49c1-4b0c-b6b7-8dcb613648fd" path="/var/lib/kubelet/pods/37eba15b-49c1-4b0c-b6b7-8dcb613648fd/volumes" Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.434218 4990 generic.go:334] "Generic (PLEG): container finished" podID="3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" containerID="0aaeac0d890447ab06b1675bbc0ecd452feed1f9651d741ea7b8a363829e68d3" exitCode=0 Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.434354 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3383-account-create-update-6ggwh" event={"ID":"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e","Type":"ContainerDied","Data":"0aaeac0d890447ab06b1675bbc0ecd452feed1f9651d741ea7b8a363829e68d3"} Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.451097 4990 generic.go:334] "Generic (PLEG): container finished" podID="a7a46743-abd7-4cc7-9594-60332bb6aa8d" containerID="8a62221fe96a9eb71de6e3730b0bc15c912092e90565f1cbfc86aed4918e7823" exitCode=0 Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.451192 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6d20-account-create-update-929g9" event={"ID":"a7a46743-abd7-4cc7-9594-60332bb6aa8d","Type":"ContainerDied","Data":"8a62221fe96a9eb71de6e3730b0bc15c912092e90565f1cbfc86aed4918e7823"} Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.454236 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbdb7170-e955-4fd4-bfd5-03ad13363aa9","Type":"ContainerStarted","Data":"866b19e22c4a2e6e9a6fad00380a888dd907a53a8554d0c922b84e8f0038fd09"} Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.454274 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbdb7170-e955-4fd4-bfd5-03ad13363aa9","Type":"ContainerStarted","Data":"7edcb2abf6c75fbdadaf8f00cd6ef33756222ad42ca9ec5e5ecd0027f7044446"} Dec 03 13:02:58 crc kubenswrapper[4990]: I1203 13:02:58.996399 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.073414 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d4ddf00-b53e-4d9c-bfaa-676226812d86-operator-scripts\") pod \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.073623 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bsrq\" (UniqueName: \"kubernetes.io/projected/8d4ddf00-b53e-4d9c-bfaa-676226812d86-kube-api-access-4bsrq\") pod \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\" (UID: \"8d4ddf00-b53e-4d9c-bfaa-676226812d86\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.074794 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d4ddf00-b53e-4d9c-bfaa-676226812d86-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d4ddf00-b53e-4d9c-bfaa-676226812d86" (UID: "8d4ddf00-b53e-4d9c-bfaa-676226812d86"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.086918 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d4ddf00-b53e-4d9c-bfaa-676226812d86-kube-api-access-4bsrq" (OuterVolumeSpecName: "kube-api-access-4bsrq") pod "8d4ddf00-b53e-4d9c-bfaa-676226812d86" (UID: "8d4ddf00-b53e-4d9c-bfaa-676226812d86"). InnerVolumeSpecName "kube-api-access-4bsrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.176324 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d4ddf00-b53e-4d9c-bfaa-676226812d86-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.176999 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bsrq\" (UniqueName: \"kubernetes.io/projected/8d4ddf00-b53e-4d9c-bfaa-676226812d86-kube-api-access-4bsrq\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.254514 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.254514 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.256741 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.335657 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.354861 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380760 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjh6j\" (UniqueName: \"kubernetes.io/projected/2aceea00-aeff-45f4-83ae-c7656a28be74-kube-api-access-vjh6j\") pod \"2aceea00-aeff-45f4-83ae-c7656a28be74\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380820 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aceea00-aeff-45f4-83ae-c7656a28be74-operator-scripts\") pod \"2aceea00-aeff-45f4-83ae-c7656a28be74\" (UID: \"2aceea00-aeff-45f4-83ae-c7656a28be74\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380870 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-internal-tls-certs\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380902 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-scripts\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380937 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-combined-ca-bundle\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380976 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b129617-7cab-45cc-9cb0-00fdf619bec2-operator-scripts\") pod \"1b129617-7cab-45cc-9cb0-00fdf619bec2\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.380996 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-config-data\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381024 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381061 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-operator-scripts\") pod \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381129 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67pbs\" (UniqueName: \"kubernetes.io/projected/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-kube-api-access-67pbs\") pod \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\" (UID: \"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381243 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-httpd-run\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381297 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmzrv\" (UniqueName: \"kubernetes.io/projected/1b129617-7cab-45cc-9cb0-00fdf619bec2-kube-api-access-rmzrv\") pod \"1b129617-7cab-45cc-9cb0-00fdf619bec2\" (UID: \"1b129617-7cab-45cc-9cb0-00fdf619bec2\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381332 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-logs\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381379 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8hzj\" (UniqueName: \"kubernetes.io/projected/1dd859cc-c778-4e79-8e61-ae6e754e21d7-kube-api-access-q8hzj\") pod \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\" (UID: \"1dd859cc-c778-4e79-8e61-ae6e754e21d7\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.381715 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b129617-7cab-45cc-9cb0-00fdf619bec2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1b129617-7cab-45cc-9cb0-00fdf619bec2" (UID: "1b129617-7cab-45cc-9cb0-00fdf619bec2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.382793 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b129617-7cab-45cc-9cb0-00fdf619bec2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.387502 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aceea00-aeff-45f4-83ae-c7656a28be74-kube-api-access-vjh6j" (OuterVolumeSpecName: "kube-api-access-vjh6j") pod "2aceea00-aeff-45f4-83ae-c7656a28be74" (UID: "2aceea00-aeff-45f4-83ae-c7656a28be74"). InnerVolumeSpecName "kube-api-access-vjh6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.387902 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aceea00-aeff-45f4-83ae-c7656a28be74-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2aceea00-aeff-45f4-83ae-c7656a28be74" (UID: "2aceea00-aeff-45f4-83ae-c7656a28be74"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.392649 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dd859cc-c778-4e79-8e61-ae6e754e21d7-kube-api-access-q8hzj" (OuterVolumeSpecName: "kube-api-access-q8hzj") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "kube-api-access-q8hzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.393601 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-logs" (OuterVolumeSpecName: "logs") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.393976 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.394505 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" (UID: "1d0fa5a6-22a0-4550-ba7e-ab7f23efad27"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.398432 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-kube-api-access-67pbs" (OuterVolumeSpecName: "kube-api-access-67pbs") pod "1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" (UID: "1d0fa5a6-22a0-4550-ba7e-ab7f23efad27"). InnerVolumeSpecName "kube-api-access-67pbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.407966 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b129617-7cab-45cc-9cb0-00fdf619bec2-kube-api-access-rmzrv" (OuterVolumeSpecName: "kube-api-access-rmzrv") pod "1b129617-7cab-45cc-9cb0-00fdf619bec2" (UID: "1b129617-7cab-45cc-9cb0-00fdf619bec2"). InnerVolumeSpecName "kube-api-access-rmzrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.410341 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.412743 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-scripts" (OuterVolumeSpecName: "scripts") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.469712 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-config-data" (OuterVolumeSpecName: "config-data") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.470152 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.475713 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-znfpn" event={"ID":"8d4ddf00-b53e-4d9c-bfaa-676226812d86","Type":"ContainerDied","Data":"d56ef90c0b8f397f600412dfb75e9082bb7788b8d5376efd39aa7617eedfde85"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.475782 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d56ef90c0b8f397f600412dfb75e9082bb7788b8d5376efd39aa7617eedfde85" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.475739 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-znfpn" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.484538 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-54jvk" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.484545 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-54jvk" event={"ID":"1d0fa5a6-22a0-4550-ba7e-ab7f23efad27","Type":"ContainerDied","Data":"340f7bfa5ad96ffd7320001276b5aefeb735a38ff585e2cd4c4ddf4c7e19cb0d"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.484597 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="340f7bfa5ad96ffd7320001276b5aefeb735a38ff585e2cd4c4ddf4c7e19cb0d" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.484657 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-run-httpd\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.484762 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9smsl\" (UniqueName: \"kubernetes.io/projected/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-kube-api-access-9smsl\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.484902 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-scripts\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.485006 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-combined-ca-bundle\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.485038 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-log-httpd\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.485082 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-config-data\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.485118 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-sg-core-conf-yaml\") pod \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\" (UID: \"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e\") " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.485195 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486852 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486885 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486903 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486935 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486949 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486962 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67pbs\" (UniqueName: \"kubernetes.io/projected/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27-kube-api-access-67pbs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486974 4990 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486987 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmzrv\" (UniqueName: \"kubernetes.io/projected/1b129617-7cab-45cc-9cb0-00fdf619bec2-kube-api-access-rmzrv\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.486999 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dd859cc-c778-4e79-8e61-ae6e754e21d7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.487010 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8hzj\" (UniqueName: \"kubernetes.io/projected/1dd859cc-c778-4e79-8e61-ae6e754e21d7-kube-api-access-q8hzj\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.487023 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjh6j\" (UniqueName: \"kubernetes.io/projected/2aceea00-aeff-45f4-83ae-c7656a28be74-kube-api-access-vjh6j\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.487034 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.487045 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aceea00-aeff-45f4-83ae-c7656a28be74-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.492041 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.492313 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1dd859cc-c778-4e79-8e61-ae6e754e21d7" (UID: "1dd859cc-c778-4e79-8e61-ae6e754e21d7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.499555 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-scripts" (OuterVolumeSpecName: "scripts") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.516163 4990 generic.go:334] "Generic (PLEG): container finished" podID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerID="9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016" exitCode=0 Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.516387 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerDied","Data":"9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.516421 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e","Type":"ContainerDied","Data":"32beec4ebbb056febeb5139b9c9e7b3a991edd2c2267322ea4e029f904a570d7"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.516466 4990 scope.go:117] "RemoveContainer" containerID="85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.516716 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.521486 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-kube-api-access-9smsl" (OuterVolumeSpecName: "kube-api-access-9smsl") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "kube-api-access-9smsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.527111 4990 generic.go:334] "Generic (PLEG): container finished" podID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerID="d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8" exitCode=0 Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.527253 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.527260 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1dd859cc-c778-4e79-8e61-ae6e754e21d7","Type":"ContainerDied","Data":"d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.527522 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1dd859cc-c778-4e79-8e61-ae6e754e21d7","Type":"ContainerDied","Data":"25ee23f1ad624163ccb404009371ec317fb793c1dd6e2725148bd4fca2ed622e"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.533810 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbdb7170-e955-4fd4-bfd5-03ad13363aa9","Type":"ContainerStarted","Data":"536e746ae8a96384bef839671359f9e9505e7bc3e95c05c6edf1c1bf9dcbdaa8"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.541632 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5857-account-create-update-bf22s" event={"ID":"1b129617-7cab-45cc-9cb0-00fdf619bec2","Type":"ContainerDied","Data":"59447b9c38b66008cf6194e759dabe8ad03d7ca0a7b0e946ce0434a664a9821c"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.541710 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59447b9c38b66008cf6194e759dabe8ad03d7ca0a7b0e946ce0434a664a9821c" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.541898 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5857-account-create-update-bf22s" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.548586 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7mtkf" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.549047 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7mtkf" event={"ID":"2aceea00-aeff-45f4-83ae-c7656a28be74","Type":"ContainerDied","Data":"abe1dec3c6cee2644327582ef0182de9ec76c421f7909b354acef343f16c050d"} Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.549084 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abe1dec3c6cee2644327582ef0182de9ec76c421f7909b354acef343f16c050d" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.554416 4990 scope.go:117] "RemoveContainer" containerID="c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.580578 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.581728 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.601549 4990 scope.go:117] "RemoveContainer" containerID="95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.604083 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.604111 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.604123 4990 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dd859cc-c778-4e79-8e61-ae6e754e21d7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.604135 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9smsl\" (UniqueName: \"kubernetes.io/projected/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-kube-api-access-9smsl\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.604147 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.604158 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.606389 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.606366302 podStartE2EDuration="3.606366302s" podCreationTimestamp="2025-12-03 13:02:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:02:59.563859942 +0000 UTC m=+1527.705771171" watchObservedRunningTime="2025-12-03 13:02:59.606366302 +0000 UTC m=+1527.748277531" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.643850 4990 scope.go:117] "RemoveContainer" containerID="9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.644057 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.644924 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.656570 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657210 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="proxy-httpd" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657228 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="proxy-httpd" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657242 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657248 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657269 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-central-agent" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657275 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-central-agent" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657283 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-notification-agent" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657292 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-notification-agent" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657307 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b129617-7cab-45cc-9cb0-00fdf619bec2" containerName="mariadb-account-create-update" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657314 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b129617-7cab-45cc-9cb0-00fdf619bec2" containerName="mariadb-account-create-update" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657330 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-httpd" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657337 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-httpd" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657348 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aceea00-aeff-45f4-83ae-c7656a28be74" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657354 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aceea00-aeff-45f4-83ae-c7656a28be74" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657365 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-log" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657370 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-log" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657383 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4ddf00-b53e-4d9c-bfaa-676226812d86" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657389 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4ddf00-b53e-4d9c-bfaa-676226812d86" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.657402 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="sg-core" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657411 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="sg-core" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657643 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-log" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657655 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b129617-7cab-45cc-9cb0-00fdf619bec2" containerName="mariadb-account-create-update" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657666 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d4ddf00-b53e-4d9c-bfaa-676226812d86" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657677 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-central-agent" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657689 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657702 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="proxy-httpd" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657714 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" containerName="glance-httpd" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657725 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aceea00-aeff-45f4-83ae-c7656a28be74" containerName="mariadb-database-create" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657738 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="sg-core" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.657745 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" containerName="ceilometer-notification-agent" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.659007 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.659131 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.667654 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.668018 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708161 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd733a23-295d-4c67-9f67-1e27d99d7847-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708222 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708322 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708443 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd733a23-295d-4c67-9f67-1e27d99d7847-logs\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708509 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708524 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708543 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8lh7\" (UniqueName: \"kubernetes.io/projected/fd733a23-295d-4c67-9f67-1e27d99d7847-kube-api-access-x8lh7\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.708605 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.722969 4990 scope.go:117] "RemoveContainer" containerID="85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.726036 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623\": container with ID starting with 85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623 not found: ID does not exist" containerID="85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.726102 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623"} err="failed to get container status \"85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623\": rpc error: code = NotFound desc = could not find container \"85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623\": container with ID starting with 85665b51904c86cecd71d1f4a09a72a790ab2aaa1cebeb9a5664133716fdb623 not found: ID does not exist" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.726148 4990 scope.go:117] "RemoveContainer" containerID="c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.727008 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63\": container with ID starting with c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63 not found: ID does not exist" containerID="c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.727031 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63"} err="failed to get container status \"c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63\": rpc error: code = NotFound desc = could not find container \"c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63\": container with ID starting with c004c5ee82ba8b42c7abc58b58edfa7411af2330f40e44598880415ee7c29c63 not found: ID does not exist" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.727050 4990 scope.go:117] "RemoveContainer" containerID="95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.727478 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30\": container with ID starting with 95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30 not found: ID does not exist" containerID="95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.727502 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30"} err="failed to get container status \"95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30\": rpc error: code = NotFound desc = could not find container \"95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30\": container with ID starting with 95eedc85597178030f06e41a121b6926626fba6e0de25a92ea1276bb75ff8a30 not found: ID does not exist" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.727523 4990 scope.go:117] "RemoveContainer" containerID="9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.727760 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016\": container with ID starting with 9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016 not found: ID does not exist" containerID="9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.727782 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016"} err="failed to get container status \"9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016\": rpc error: code = NotFound desc = could not find container \"9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016\": container with ID starting with 9f597222f33c8f6f16898b0d930548cb46ca336186a1ef92d21af6ebda7ba016 not found: ID does not exist" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.727800 4990 scope.go:117] "RemoveContainer" containerID="d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.757192 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.776979 4990 scope.go:117] "RemoveContainer" containerID="0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.778287 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-config-data" (OuterVolumeSpecName: "config-data") pod "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" (UID: "ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810135 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810213 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810247 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8lh7\" (UniqueName: \"kubernetes.io/projected/fd733a23-295d-4c67-9f67-1e27d99d7847-kube-api-access-x8lh7\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810300 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810354 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd733a23-295d-4c67-9f67-1e27d99d7847-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810385 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810468 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810549 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd733a23-295d-4c67-9f67-1e27d99d7847-logs\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810620 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.810634 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.811204 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd733a23-295d-4c67-9f67-1e27d99d7847-logs\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.811383 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.812204 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd733a23-295d-4c67-9f67-1e27d99d7847-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.815679 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.816064 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.823567 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.828620 4990 scope.go:117] "RemoveContainer" containerID="d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.829222 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8\": container with ID starting with d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8 not found: ID does not exist" containerID="d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.829356 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8"} err="failed to get container status \"d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8\": rpc error: code = NotFound desc = could not find container \"d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8\": container with ID starting with d306d0000c35b4633013f4e1d320604b9f5f2ce84bdc98368ee172dc5d9a2fe8 not found: ID does not exist" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.829527 4990 scope.go:117] "RemoveContainer" containerID="0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409" Dec 03 13:02:59 crc kubenswrapper[4990]: E1203 13:02:59.830161 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409\": container with ID starting with 0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409 not found: ID does not exist" containerID="0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.830224 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409"} err="failed to get container status \"0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409\": rpc error: code = NotFound desc = could not find container \"0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409\": container with ID starting with 0e203143b20819762d2580512ff9e41befba1de025ee806a3b7fe2fcb0dd2409 not found: ID does not exist" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.831936 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8lh7\" (UniqueName: \"kubernetes.io/projected/fd733a23-295d-4c67-9f67-1e27d99d7847-kube-api-access-x8lh7\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.840827 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd733a23-295d-4c67-9f67-1e27d99d7847-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.845957 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd733a23-295d-4c67-9f67-1e27d99d7847\") " pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.886042 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.910240 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.920308 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.922536 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.928872 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.930275 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.933493 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.991321 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 13:02:59 crc kubenswrapper[4990]: I1203 13:02:59.993708 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.022376 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a46743-abd7-4cc7-9594-60332bb6aa8d-operator-scripts\") pod \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.022579 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnxmv\" (UniqueName: \"kubernetes.io/projected/a7a46743-abd7-4cc7-9594-60332bb6aa8d-kube-api-access-wnxmv\") pod \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\" (UID: \"a7a46743-abd7-4cc7-9594-60332bb6aa8d\") " Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.022886 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-scripts\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.022912 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.022959 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tq4v\" (UniqueName: \"kubernetes.io/projected/926f7e1a-309e-4fc1-aa99-84db233a97ff-kube-api-access-5tq4v\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.022983 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-config-data\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.023016 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7a46743-abd7-4cc7-9594-60332bb6aa8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7a46743-abd7-4cc7-9594-60332bb6aa8d" (UID: "a7a46743-abd7-4cc7-9594-60332bb6aa8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.023187 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-run-httpd\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.023277 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.023348 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-log-httpd\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.023832 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7a46743-abd7-4cc7-9594-60332bb6aa8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.030650 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a46743-abd7-4cc7-9594-60332bb6aa8d-kube-api-access-wnxmv" (OuterVolumeSpecName: "kube-api-access-wnxmv") pod "a7a46743-abd7-4cc7-9594-60332bb6aa8d" (UID: "a7a46743-abd7-4cc7-9594-60332bb6aa8d"). InnerVolumeSpecName "kube-api-access-wnxmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.126975 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-log-httpd\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127416 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-scripts\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127442 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127550 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tq4v\" (UniqueName: \"kubernetes.io/projected/926f7e1a-309e-4fc1-aa99-84db233a97ff-kube-api-access-5tq4v\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127578 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-config-data\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127669 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-run-httpd\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127572 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-log-httpd\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127719 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.127990 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnxmv\" (UniqueName: \"kubernetes.io/projected/a7a46743-abd7-4cc7-9594-60332bb6aa8d-kube-api-access-wnxmv\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.128280 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-run-httpd\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.136660 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.136927 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-scripts\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.137806 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.146304 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-config-data\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.148042 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tq4v\" (UniqueName: \"kubernetes.io/projected/926f7e1a-309e-4fc1-aa99-84db233a97ff-kube-api-access-5tq4v\") pod \"ceilometer-0\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.217963 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.246059 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.284007 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dd859cc-c778-4e79-8e61-ae6e754e21d7" path="/var/lib/kubelet/pods/1dd859cc-c778-4e79-8e61-ae6e754e21d7/volumes" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.285273 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e" path="/var/lib/kubelet/pods/ec0330c3-01e2-4a52-9a7b-5b1d467cbe9e/volumes" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.330474 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khfkh\" (UniqueName: \"kubernetes.io/projected/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-kube-api-access-khfkh\") pod \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.331439 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" (UID: "3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.331505 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-operator-scripts\") pod \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\" (UID: \"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e\") " Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.333096 4990 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.344782 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-kube-api-access-khfkh" (OuterVolumeSpecName: "kube-api-access-khfkh") pod "3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" (UID: "3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e"). InnerVolumeSpecName "kube-api-access-khfkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.435513 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khfkh\" (UniqueName: \"kubernetes.io/projected/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e-kube-api-access-khfkh\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.562830 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3383-account-create-update-6ggwh" event={"ID":"3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e","Type":"ContainerDied","Data":"6a5dba92e88ef4defb613815cd4b1ad1bf2cda4b05faf4ff48f916b26743acaf"} Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.562881 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a5dba92e88ef4defb613815cd4b1ad1bf2cda4b05faf4ff48f916b26743acaf" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.562846 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3383-account-create-update-6ggwh" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.564420 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6d20-account-create-update-929g9" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.564444 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6d20-account-create-update-929g9" event={"ID":"a7a46743-abd7-4cc7-9594-60332bb6aa8d","Type":"ContainerDied","Data":"6c0dfcdabb4d76e4f9b9c7f26d470885478919d53903cd4f91455c230a7f1762"} Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.564506 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c0dfcdabb4d76e4f9b9c7f26d470885478919d53903cd4f91455c230a7f1762" Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.586116 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 13:03:00 crc kubenswrapper[4990]: I1203 13:03:00.714707 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:00 crc kubenswrapper[4990]: W1203 13:03:00.724355 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod926f7e1a_309e_4fc1_aa99_84db233a97ff.slice/crio-803aa5f2cfde2b0218ec8d8bedec3ed940ca48fab95c004dd96cd6174ffe771f WatchSource:0}: Error finding container 803aa5f2cfde2b0218ec8d8bedec3ed940ca48fab95c004dd96cd6174ffe771f: Status 404 returned error can't find the container with id 803aa5f2cfde2b0218ec8d8bedec3ed940ca48fab95c004dd96cd6174ffe771f Dec 03 13:03:01 crc kubenswrapper[4990]: I1203 13:03:01.592119 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerStarted","Data":"803aa5f2cfde2b0218ec8d8bedec3ed940ca48fab95c004dd96cd6174ffe771f"} Dec 03 13:03:01 crc kubenswrapper[4990]: I1203 13:03:01.596882 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd733a23-295d-4c67-9f67-1e27d99d7847","Type":"ContainerStarted","Data":"e86f27192286fc4a3a37234d7f55d67f777ad0a337c0868e13f1bd56d795094f"} Dec 03 13:03:01 crc kubenswrapper[4990]: I1203 13:03:01.596962 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd733a23-295d-4c67-9f67-1e27d99d7847","Type":"ContainerStarted","Data":"8c40585a5e0cdfabd0e7948cf9455308ee18d144394401c3c289aadfa41f3a4b"} Dec 03 13:03:02 crc kubenswrapper[4990]: I1203 13:03:02.574250 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:02 crc kubenswrapper[4990]: I1203 13:03:02.608092 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerStarted","Data":"a7f5a80afefc8455f07be3b36fd86d3ed2e945e03537c62155a3a92aeefb9684"} Dec 03 13:03:02 crc kubenswrapper[4990]: I1203 13:03:02.613468 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd733a23-295d-4c67-9f67-1e27d99d7847","Type":"ContainerStarted","Data":"99ad6833b6d08f61247c81e22fd30c8b637269b4ee8ebc0e3e8ebe2f28f484e3"} Dec 03 13:03:02 crc kubenswrapper[4990]: I1203 13:03:02.633133 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.633113269 podStartE2EDuration="3.633113269s" podCreationTimestamp="2025-12-03 13:02:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:03:02.630805338 +0000 UTC m=+1530.772716567" watchObservedRunningTime="2025-12-03 13:03:02.633113269 +0000 UTC m=+1530.775024498" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.178205 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vkq7n"] Dec 03 13:03:03 crc kubenswrapper[4990]: E1203 13:03:03.178648 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" containerName="mariadb-account-create-update" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.178673 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" containerName="mariadb-account-create-update" Dec 03 13:03:03 crc kubenswrapper[4990]: E1203 13:03:03.178716 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a46743-abd7-4cc7-9594-60332bb6aa8d" containerName="mariadb-account-create-update" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.178725 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a46743-abd7-4cc7-9594-60332bb6aa8d" containerName="mariadb-account-create-update" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.179580 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a46743-abd7-4cc7-9594-60332bb6aa8d" containerName="mariadb-account-create-update" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.179695 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" containerName="mariadb-account-create-update" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.180360 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.182717 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5dljg" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.182996 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.183465 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.188297 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vkq7n"] Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.287385 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.287462 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.300594 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.300729 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-config-data\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.300754 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-scripts\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.300815 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbv8x\" (UniqueName: \"kubernetes.io/projected/9f6005f7-25b3-4e17-bacf-763d1f68e22c-kube-api-access-xbv8x\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.402519 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-config-data\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.402567 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-scripts\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.402617 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbv8x\" (UniqueName: \"kubernetes.io/projected/9f6005f7-25b3-4e17-bacf-763d1f68e22c-kube-api-access-xbv8x\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.402667 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.408233 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-scripts\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.410010 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.410103 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-config-data\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.421633 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbv8x\" (UniqueName: \"kubernetes.io/projected/9f6005f7-25b3-4e17-bacf-763d1f68e22c-kube-api-access-xbv8x\") pod \"nova-cell0-conductor-db-sync-vkq7n\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.504755 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:03 crc kubenswrapper[4990]: I1203 13:03:03.621765 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerStarted","Data":"1ecb28766307600c957f029322252c379b7d44956ee690249312ced17fae9b96"} Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.176791 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vkq7n"] Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.659653 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" event={"ID":"9f6005f7-25b3-4e17-bacf-763d1f68e22c","Type":"ContainerStarted","Data":"3ae5b3b53654e7c755853231d4010a134d1f5b9aa0f28b3489d9f0d28d191499"} Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.666971 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerStarted","Data":"c25e655e9296dc517750e949e4fc957c8cd67d10702539de8b37d6ddccf1bfad"} Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.788597 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.790297 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.827519 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:03:06 crc kubenswrapper[4990]: I1203 13:03:06.847997 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 13:03:07 crc kubenswrapper[4990]: I1203 13:03:07.676568 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:03:07 crc kubenswrapper[4990]: I1203 13:03:07.677024 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.702744 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerStarted","Data":"7c0eaf86f57a0d15f1a91ff0d0bf4b8cec88f88e29104bdd6ea35d58a695bb8c"} Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.702768 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.703124 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.702883 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-central-agent" containerID="cri-o://a7f5a80afefc8455f07be3b36fd86d3ed2e945e03537c62155a3a92aeefb9684" gracePeriod=30 Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.702964 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="proxy-httpd" containerID="cri-o://7c0eaf86f57a0d15f1a91ff0d0bf4b8cec88f88e29104bdd6ea35d58a695bb8c" gracePeriod=30 Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.702935 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="sg-core" containerID="cri-o://c25e655e9296dc517750e949e4fc957c8cd67d10702539de8b37d6ddccf1bfad" gracePeriod=30 Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.702975 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-notification-agent" containerID="cri-o://1ecb28766307600c957f029322252c379b7d44956ee690249312ced17fae9b96" gracePeriod=30 Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.703363 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.738545 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.638256465 podStartE2EDuration="10.73852607s" podCreationTimestamp="2025-12-03 13:02:59 +0000 UTC" firstStartedPulling="2025-12-03 13:03:00.727996651 +0000 UTC m=+1528.869907880" lastFinishedPulling="2025-12-03 13:03:08.828266256 +0000 UTC m=+1536.970177485" observedRunningTime="2025-12-03 13:03:09.731198827 +0000 UTC m=+1537.873110076" watchObservedRunningTime="2025-12-03 13:03:09.73852607 +0000 UTC m=+1537.880437299" Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.762021 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:03:09 crc kubenswrapper[4990]: I1203 13:03:09.959987 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:09.994706 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:09.996215 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.094245 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.096074 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714203 4990 generic.go:334] "Generic (PLEG): container finished" podID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerID="7c0eaf86f57a0d15f1a91ff0d0bf4b8cec88f88e29104bdd6ea35d58a695bb8c" exitCode=0 Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714241 4990 generic.go:334] "Generic (PLEG): container finished" podID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerID="c25e655e9296dc517750e949e4fc957c8cd67d10702539de8b37d6ddccf1bfad" exitCode=2 Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714249 4990 generic.go:334] "Generic (PLEG): container finished" podID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerID="1ecb28766307600c957f029322252c379b7d44956ee690249312ced17fae9b96" exitCode=0 Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714260 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerDied","Data":"7c0eaf86f57a0d15f1a91ff0d0bf4b8cec88f88e29104bdd6ea35d58a695bb8c"} Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714320 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerDied","Data":"c25e655e9296dc517750e949e4fc957c8cd67d10702539de8b37d6ddccf1bfad"} Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714334 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerDied","Data":"1ecb28766307600c957f029322252c379b7d44956ee690249312ced17fae9b96"} Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714772 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:10 crc kubenswrapper[4990]: I1203 13:03:10.714793 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:12 crc kubenswrapper[4990]: I1203 13:03:12.758601 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:12 crc kubenswrapper[4990]: I1203 13:03:12.759626 4990 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 13:03:12 crc kubenswrapper[4990]: I1203 13:03:12.902632 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 13:03:13 crc kubenswrapper[4990]: I1203 13:03:13.756363 4990 generic.go:334] "Generic (PLEG): container finished" podID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerID="a7f5a80afefc8455f07be3b36fd86d3ed2e945e03537c62155a3a92aeefb9684" exitCode=0 Dec 03 13:03:13 crc kubenswrapper[4990]: I1203 13:03:13.756439 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerDied","Data":"a7f5a80afefc8455f07be3b36fd86d3ed2e945e03537c62155a3a92aeefb9684"} Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.260282 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.328903 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-config-data\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.329003 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-log-httpd\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.329049 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-sg-core-conf-yaml\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.329179 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-combined-ca-bundle\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.329232 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-run-httpd\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.329264 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-scripts\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.329293 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tq4v\" (UniqueName: \"kubernetes.io/projected/926f7e1a-309e-4fc1-aa99-84db233a97ff-kube-api-access-5tq4v\") pod \"926f7e1a-309e-4fc1-aa99-84db233a97ff\" (UID: \"926f7e1a-309e-4fc1-aa99-84db233a97ff\") " Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.330771 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.331651 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.334614 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/926f7e1a-309e-4fc1-aa99-84db233a97ff-kube-api-access-5tq4v" (OuterVolumeSpecName: "kube-api-access-5tq4v") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "kube-api-access-5tq4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.337655 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-scripts" (OuterVolumeSpecName: "scripts") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.360673 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.413576 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.431394 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.431428 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.431437 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tq4v\" (UniqueName: \"kubernetes.io/projected/926f7e1a-309e-4fc1-aa99-84db233a97ff-kube-api-access-5tq4v\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.431451 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/926f7e1a-309e-4fc1-aa99-84db233a97ff-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.431459 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.431470 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.435976 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-config-data" (OuterVolumeSpecName: "config-data") pod "926f7e1a-309e-4fc1-aa99-84db233a97ff" (UID: "926f7e1a-309e-4fc1-aa99-84db233a97ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.532996 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f7e1a-309e-4fc1-aa99-84db233a97ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.823458 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" event={"ID":"9f6005f7-25b3-4e17-bacf-763d1f68e22c","Type":"ContainerStarted","Data":"133b18a07e1097b63c89fec7ad14818a5236027a276aca7e8f19f8e29508fa23"} Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.827670 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"926f7e1a-309e-4fc1-aa99-84db233a97ff","Type":"ContainerDied","Data":"803aa5f2cfde2b0218ec8d8bedec3ed940ca48fab95c004dd96cd6174ffe771f"} Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.827728 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.827736 4990 scope.go:117] "RemoveContainer" containerID="7c0eaf86f57a0d15f1a91ff0d0bf4b8cec88f88e29104bdd6ea35d58a695bb8c" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.848903 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" podStartSLOduration=4.027392277 podStartE2EDuration="16.848872992s" podCreationTimestamp="2025-12-03 13:03:03 +0000 UTC" firstStartedPulling="2025-12-03 13:03:06.1837436 +0000 UTC m=+1534.325654849" lastFinishedPulling="2025-12-03 13:03:19.005224335 +0000 UTC m=+1547.147135564" observedRunningTime="2025-12-03 13:03:19.843645334 +0000 UTC m=+1547.985556573" watchObservedRunningTime="2025-12-03 13:03:19.848872992 +0000 UTC m=+1547.990784221" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.877635 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.878531 4990 scope.go:117] "RemoveContainer" containerID="c25e655e9296dc517750e949e4fc957c8cd67d10702539de8b37d6ddccf1bfad" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.894995 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.907432 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:19 crc kubenswrapper[4990]: E1203 13:03:19.907897 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-notification-agent" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.907913 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-notification-agent" Dec 03 13:03:19 crc kubenswrapper[4990]: E1203 13:03:19.907933 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="sg-core" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.907939 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="sg-core" Dec 03 13:03:19 crc kubenswrapper[4990]: E1203 13:03:19.907965 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-central-agent" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.907971 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-central-agent" Dec 03 13:03:19 crc kubenswrapper[4990]: E1203 13:03:19.907986 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="proxy-httpd" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.907992 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="proxy-httpd" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.908147 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-notification-agent" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.908163 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="ceilometer-central-agent" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.908177 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="proxy-httpd" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.908189 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" containerName="sg-core" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.909809 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.912878 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.915938 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.924045 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.937854 4990 scope.go:117] "RemoveContainer" containerID="1ecb28766307600c957f029322252c379b7d44956ee690249312ced17fae9b96" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943438 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943563 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-run-httpd\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943625 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-scripts\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943659 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-config-data\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943783 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwpw6\" (UniqueName: \"kubernetes.io/projected/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-kube-api-access-cwpw6\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943820 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-log-httpd\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.943929 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:19 crc kubenswrapper[4990]: I1203 13:03:19.968209 4990 scope.go:117] "RemoveContainer" containerID="a7f5a80afefc8455f07be3b36fd86d3ed2e945e03537c62155a3a92aeefb9684" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045463 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045566 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045635 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-run-httpd\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045697 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-scripts\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045739 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-config-data\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045781 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwpw6\" (UniqueName: \"kubernetes.io/projected/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-kube-api-access-cwpw6\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.045806 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-log-httpd\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.046255 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-run-httpd\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.046270 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-log-httpd\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.050653 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-config-data\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.050851 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.050986 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-scripts\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.057978 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.075358 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwpw6\" (UniqueName: \"kubernetes.io/projected/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-kube-api-access-cwpw6\") pod \"ceilometer-0\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.230287 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.283826 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="926f7e1a-309e-4fc1-aa99-84db233a97ff" path="/var/lib/kubelet/pods/926f7e1a-309e-4fc1-aa99-84db233a97ff/volumes" Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.695763 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:20 crc kubenswrapper[4990]: W1203 13:03:20.697046 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f5c7c4_0f73_4670_bbdc_ebafc6b36a01.slice/crio-266c3af62f66e4a99cd5989e431aa226b7921ce3a7778f455140fac2704b597c WatchSource:0}: Error finding container 266c3af62f66e4a99cd5989e431aa226b7921ce3a7778f455140fac2704b597c: Status 404 returned error can't find the container with id 266c3af62f66e4a99cd5989e431aa226b7921ce3a7778f455140fac2704b597c Dec 03 13:03:20 crc kubenswrapper[4990]: I1203 13:03:20.839258 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerStarted","Data":"266c3af62f66e4a99cd5989e431aa226b7921ce3a7778f455140fac2704b597c"} Dec 03 13:03:21 crc kubenswrapper[4990]: I1203 13:03:21.334917 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:26 crc kubenswrapper[4990]: I1203 13:03:26.910645 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerStarted","Data":"d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3"} Dec 03 13:03:27 crc kubenswrapper[4990]: I1203 13:03:27.921498 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerStarted","Data":"857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722"} Dec 03 13:03:30 crc kubenswrapper[4990]: I1203 13:03:30.956325 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerStarted","Data":"9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3"} Dec 03 13:03:33 crc kubenswrapper[4990]: I1203 13:03:33.286597 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:03:33 crc kubenswrapper[4990]: I1203 13:03:33.287219 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:03:33 crc kubenswrapper[4990]: I1203 13:03:33.287967 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:03:33 crc kubenswrapper[4990]: I1203 13:03:33.289766 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:03:33 crc kubenswrapper[4990]: I1203 13:03:33.289902 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" gracePeriod=600 Dec 03 13:03:34 crc kubenswrapper[4990]: I1203 13:03:34.000533 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" exitCode=0 Dec 03 13:03:34 crc kubenswrapper[4990]: I1203 13:03:34.000727 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b"} Dec 03 13:03:34 crc kubenswrapper[4990]: I1203 13:03:34.001073 4990 scope.go:117] "RemoveContainer" containerID="af393e5087a271fce924a6dbdf31ad1b2b0bea5f1a2211228838f21a6a0541a1" Dec 03 13:03:34 crc kubenswrapper[4990]: E1203 13:03:34.705170 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:03:35 crc kubenswrapper[4990]: I1203 13:03:35.009676 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:03:35 crc kubenswrapper[4990]: E1203 13:03:35.010000 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.050002 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerStarted","Data":"ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63"} Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.050329 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-central-agent" containerID="cri-o://d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3" gracePeriod=30 Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.050612 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-notification-agent" containerID="cri-o://857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722" gracePeriod=30 Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.050447 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="sg-core" containerID="cri-o://9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3" gracePeriod=30 Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.050552 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="proxy-httpd" containerID="cri-o://ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63" gracePeriod=30 Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.050818 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:03:38 crc kubenswrapper[4990]: I1203 13:03:38.083427 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.655797501 podStartE2EDuration="19.08339473s" podCreationTimestamp="2025-12-03 13:03:19 +0000 UTC" firstStartedPulling="2025-12-03 13:03:20.700248593 +0000 UTC m=+1548.842159822" lastFinishedPulling="2025-12-03 13:03:36.127845832 +0000 UTC m=+1564.269757051" observedRunningTime="2025-12-03 13:03:38.081108641 +0000 UTC m=+1566.223019910" watchObservedRunningTime="2025-12-03 13:03:38.08339473 +0000 UTC m=+1566.225305999" Dec 03 13:03:38 crc kubenswrapper[4990]: E1203 13:03:38.518090 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f5c7c4_0f73_4670_bbdc_ebafc6b36a01.slice/crio-conmon-ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f5c7c4_0f73_4670_bbdc_ebafc6b36a01.slice/crio-conmon-9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.062828 4990 generic.go:334] "Generic (PLEG): container finished" podID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerID="ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63" exitCode=0 Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.063190 4990 generic.go:334] "Generic (PLEG): container finished" podID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerID="9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3" exitCode=2 Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.063203 4990 generic.go:334] "Generic (PLEG): container finished" podID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerID="857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722" exitCode=0 Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.062904 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerDied","Data":"ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63"} Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.063241 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerDied","Data":"9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3"} Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.063257 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerDied","Data":"857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722"} Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.664328 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789147 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-scripts\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789270 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-config-data\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789482 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-sg-core-conf-yaml\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789530 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwpw6\" (UniqueName: \"kubernetes.io/projected/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-kube-api-access-cwpw6\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789570 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-log-httpd\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789636 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-combined-ca-bundle\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.789789 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-run-httpd\") pod \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\" (UID: \"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01\") " Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.790320 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.790785 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.790913 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.794442 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-kube-api-access-cwpw6" (OuterVolumeSpecName: "kube-api-access-cwpw6") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "kube-api-access-cwpw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.795836 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-scripts" (OuterVolumeSpecName: "scripts") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.814172 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.854887 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.885431 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-config-data" (OuterVolumeSpecName: "config-data") pod "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" (UID: "95f5c7c4-0f73-4670-bbdc-ebafc6b36a01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.892599 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.892633 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.892643 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.892654 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwpw6\" (UniqueName: \"kubernetes.io/projected/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-kube-api-access-cwpw6\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.892663 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:39 crc kubenswrapper[4990]: I1203 13:03:39.892671 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.075429 4990 generic.go:334] "Generic (PLEG): container finished" podID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerID="d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3" exitCode=0 Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.075511 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerDied","Data":"d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3"} Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.075553 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95f5c7c4-0f73-4670-bbdc-ebafc6b36a01","Type":"ContainerDied","Data":"266c3af62f66e4a99cd5989e431aa226b7921ce3a7778f455140fac2704b597c"} Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.075586 4990 scope.go:117] "RemoveContainer" containerID="ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.075780 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.094636 4990 scope.go:117] "RemoveContainer" containerID="9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.124967 4990 scope.go:117] "RemoveContainer" containerID="857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.131556 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.136866 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.149737 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.150112 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="sg-core" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150128 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="sg-core" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.150140 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="proxy-httpd" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150145 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="proxy-httpd" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.150162 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-notification-agent" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150168 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-notification-agent" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.150184 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-central-agent" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150190 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-central-agent" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150354 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="proxy-httpd" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150372 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-central-agent" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150381 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="ceilometer-notification-agent" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.150394 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" containerName="sg-core" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.151977 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.155893 4990 scope.go:117] "RemoveContainer" containerID="d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.156971 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.157064 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.172883 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.203904 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-scripts\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.203983 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-run-httpd\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.204258 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.204333 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-config-data\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.204608 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-log-httpd\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.204734 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-597lq\" (UniqueName: \"kubernetes.io/projected/17e5d538-2a29-4f9d-82b3-ba51790e792a-kube-api-access-597lq\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.204840 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.225676 4990 scope.go:117] "RemoveContainer" containerID="ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.226095 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63\": container with ID starting with ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63 not found: ID does not exist" containerID="ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.226147 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63"} err="failed to get container status \"ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63\": rpc error: code = NotFound desc = could not find container \"ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63\": container with ID starting with ce496faeb419677ef591b22386e221c5576fba8b0ffb858444bded41f7918b63 not found: ID does not exist" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.226178 4990 scope.go:117] "RemoveContainer" containerID="9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.226745 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3\": container with ID starting with 9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3 not found: ID does not exist" containerID="9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.226791 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3"} err="failed to get container status \"9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3\": rpc error: code = NotFound desc = could not find container \"9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3\": container with ID starting with 9829ca2d78290e1b43905a695b0926ce67c82e1bd42b017850f30dfa6bcbf6b3 not found: ID does not exist" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.226819 4990 scope.go:117] "RemoveContainer" containerID="857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.227174 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722\": container with ID starting with 857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722 not found: ID does not exist" containerID="857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.227198 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722"} err="failed to get container status \"857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722\": rpc error: code = NotFound desc = could not find container \"857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722\": container with ID starting with 857aa343c4b2b1f8a73dd3c6455e5e7834228c6eba44f07f3634a58d49b16722 not found: ID does not exist" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.227213 4990 scope.go:117] "RemoveContainer" containerID="d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3" Dec 03 13:03:40 crc kubenswrapper[4990]: E1203 13:03:40.227468 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3\": container with ID starting with d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3 not found: ID does not exist" containerID="d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.227502 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3"} err="failed to get container status \"d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3\": rpc error: code = NotFound desc = could not find container \"d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3\": container with ID starting with d247a216e59e8ce7f23b438e53ff9d676cbcf1bd15bb5ab523be9d564cce7ff3 not found: ID does not exist" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.275489 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f5c7c4-0f73-4670-bbdc-ebafc6b36a01" path="/var/lib/kubelet/pods/95f5c7c4-0f73-4670-bbdc-ebafc6b36a01/volumes" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.307379 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-scripts\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.307510 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-run-httpd\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.307662 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.307722 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-config-data\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.307773 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-log-httpd\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.308375 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-597lq\" (UniqueName: \"kubernetes.io/projected/17e5d538-2a29-4f9d-82b3-ba51790e792a-kube-api-access-597lq\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.308414 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-run-httpd\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.308656 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.308786 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-log-httpd\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.313607 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.314032 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.314984 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-config-data\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.315119 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-scripts\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.334627 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-597lq\" (UniqueName: \"kubernetes.io/projected/17e5d538-2a29-4f9d-82b3-ba51790e792a-kube-api-access-597lq\") pod \"ceilometer-0\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.528522 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:03:40 crc kubenswrapper[4990]: I1203 13:03:40.983972 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:03:41 crc kubenswrapper[4990]: I1203 13:03:41.090115 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerStarted","Data":"85bfdeb2105685d680ea7e39cd4ce8d3f9736227fa023fa34fd41572145017ed"} Dec 03 13:03:43 crc kubenswrapper[4990]: I1203 13:03:43.121259 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerStarted","Data":"e809a89e9f94386ae715b68d217fff48ea377ef7a54f73a7c5f830b24faeb48f"} Dec 03 13:03:44 crc kubenswrapper[4990]: I1203 13:03:44.131120 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerStarted","Data":"c6c94129175c04aeb0851b0e62bbeaf17cac646b9290dc8c11546d2620988ee1"} Dec 03 13:03:44 crc kubenswrapper[4990]: I1203 13:03:44.131668 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerStarted","Data":"6d04fa05fe827e2b4dd9e63743d094d0f5a21a64ae5405ef20dd30510790e2f5"} Dec 03 13:03:46 crc kubenswrapper[4990]: I1203 13:03:46.268882 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:03:46 crc kubenswrapper[4990]: E1203 13:03:46.269414 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:03:47 crc kubenswrapper[4990]: I1203 13:03:47.166608 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerStarted","Data":"6076acc03eb36ee3c3e352f49f74c6f47311211cc7734e52d0ec1e11bd96629c"} Dec 03 13:03:47 crc kubenswrapper[4990]: I1203 13:03:47.166950 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:03:47 crc kubenswrapper[4990]: I1203 13:03:47.195013 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8448405509999999 podStartE2EDuration="7.194989555s" podCreationTimestamp="2025-12-03 13:03:40 +0000 UTC" firstStartedPulling="2025-12-03 13:03:40.981050953 +0000 UTC m=+1569.122962182" lastFinishedPulling="2025-12-03 13:03:46.331199957 +0000 UTC m=+1574.473111186" observedRunningTime="2025-12-03 13:03:47.18910495 +0000 UTC m=+1575.331016179" watchObservedRunningTime="2025-12-03 13:03:47.194989555 +0000 UTC m=+1575.336900794" Dec 03 13:03:55 crc kubenswrapper[4990]: I1203 13:03:55.256104 4990 generic.go:334] "Generic (PLEG): container finished" podID="9f6005f7-25b3-4e17-bacf-763d1f68e22c" containerID="133b18a07e1097b63c89fec7ad14818a5236027a276aca7e8f19f8e29508fa23" exitCode=0 Dec 03 13:03:55 crc kubenswrapper[4990]: I1203 13:03:55.256190 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" event={"ID":"9f6005f7-25b3-4e17-bacf-763d1f68e22c","Type":"ContainerDied","Data":"133b18a07e1097b63c89fec7ad14818a5236027a276aca7e8f19f8e29508fa23"} Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.605350 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.749193 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbv8x\" (UniqueName: \"kubernetes.io/projected/9f6005f7-25b3-4e17-bacf-763d1f68e22c-kube-api-access-xbv8x\") pod \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.749327 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-combined-ca-bundle\") pod \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.749363 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-scripts\") pod \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.749393 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-config-data\") pod \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\" (UID: \"9f6005f7-25b3-4e17-bacf-763d1f68e22c\") " Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.756241 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f6005f7-25b3-4e17-bacf-763d1f68e22c-kube-api-access-xbv8x" (OuterVolumeSpecName: "kube-api-access-xbv8x") pod "9f6005f7-25b3-4e17-bacf-763d1f68e22c" (UID: "9f6005f7-25b3-4e17-bacf-763d1f68e22c"). InnerVolumeSpecName "kube-api-access-xbv8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.759588 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-scripts" (OuterVolumeSpecName: "scripts") pod "9f6005f7-25b3-4e17-bacf-763d1f68e22c" (UID: "9f6005f7-25b3-4e17-bacf-763d1f68e22c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.779324 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-config-data" (OuterVolumeSpecName: "config-data") pod "9f6005f7-25b3-4e17-bacf-763d1f68e22c" (UID: "9f6005f7-25b3-4e17-bacf-763d1f68e22c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.781652 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f6005f7-25b3-4e17-bacf-763d1f68e22c" (UID: "9f6005f7-25b3-4e17-bacf-763d1f68e22c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.851066 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbv8x\" (UniqueName: \"kubernetes.io/projected/9f6005f7-25b3-4e17-bacf-763d1f68e22c-kube-api-access-xbv8x\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.851096 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.851105 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:56 crc kubenswrapper[4990]: I1203 13:03:56.851114 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6005f7-25b3-4e17-bacf-763d1f68e22c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.277642 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" event={"ID":"9f6005f7-25b3-4e17-bacf-763d1f68e22c","Type":"ContainerDied","Data":"3ae5b3b53654e7c755853231d4010a134d1f5b9aa0f28b3489d9f0d28d191499"} Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.278258 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ae5b3b53654e7c755853231d4010a134d1f5b9aa0f28b3489d9f0d28d191499" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.277694 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vkq7n" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.374855 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 13:03:57 crc kubenswrapper[4990]: E1203 13:03:57.375376 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f6005f7-25b3-4e17-bacf-763d1f68e22c" containerName="nova-cell0-conductor-db-sync" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.375558 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f6005f7-25b3-4e17-bacf-763d1f68e22c" containerName="nova-cell0-conductor-db-sync" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.375821 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f6005f7-25b3-4e17-bacf-763d1f68e22c" containerName="nova-cell0-conductor-db-sync" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.376639 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.386693 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.391740 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-5dljg" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.399248 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.565645 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2xp8\" (UniqueName: \"kubernetes.io/projected/11fb2cd9-0846-413e-b00a-29f258f641c5-kube-api-access-p2xp8\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.565708 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11fb2cd9-0846-413e-b00a-29f258f641c5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.565938 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fb2cd9-0846-413e-b00a-29f258f641c5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.667550 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11fb2cd9-0846-413e-b00a-29f258f641c5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.668312 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fb2cd9-0846-413e-b00a-29f258f641c5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.668544 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2xp8\" (UniqueName: \"kubernetes.io/projected/11fb2cd9-0846-413e-b00a-29f258f641c5-kube-api-access-p2xp8\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.671812 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11fb2cd9-0846-413e-b00a-29f258f641c5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.673013 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11fb2cd9-0846-413e-b00a-29f258f641c5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.686623 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2xp8\" (UniqueName: \"kubernetes.io/projected/11fb2cd9-0846-413e-b00a-29f258f641c5-kube-api-access-p2xp8\") pod \"nova-cell0-conductor-0\" (UID: \"11fb2cd9-0846-413e-b00a-29f258f641c5\") " pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:57 crc kubenswrapper[4990]: I1203 13:03:57.702856 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:58 crc kubenswrapper[4990]: I1203 13:03:58.169646 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 13:03:58 crc kubenswrapper[4990]: I1203 13:03:58.264893 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:03:58 crc kubenswrapper[4990]: E1203 13:03:58.265282 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:03:58 crc kubenswrapper[4990]: I1203 13:03:58.295605 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"11fb2cd9-0846-413e-b00a-29f258f641c5","Type":"ContainerStarted","Data":"206ed3ad7b45a68c17336e62d47fde8808d08f4676d3897d795c0a156a4cf861"} Dec 03 13:03:59 crc kubenswrapper[4990]: I1203 13:03:59.305213 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"11fb2cd9-0846-413e-b00a-29f258f641c5","Type":"ContainerStarted","Data":"d693f8e701f1cc92d4d70168b3ad8ee04531ac51304fa6b8461fe6cf67146dff"} Dec 03 13:03:59 crc kubenswrapper[4990]: I1203 13:03:59.306197 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 13:03:59 crc kubenswrapper[4990]: I1203 13:03:59.329512 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.329492824 podStartE2EDuration="2.329492824s" podCreationTimestamp="2025-12-03 13:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:03:59.320631201 +0000 UTC m=+1587.462542430" watchObservedRunningTime="2025-12-03 13:03:59.329492824 +0000 UTC m=+1587.471404063" Dec 03 13:04:07 crc kubenswrapper[4990]: I1203 13:04:07.731011 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.193961 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5szpf"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.195964 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.198731 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.198732 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.202921 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5szpf"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.274927 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd4zs\" (UniqueName: \"kubernetes.io/projected/89eeb481-64d0-4b5a-b470-8edd0091c19b-kube-api-access-nd4zs\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.275231 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.275344 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-config-data\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.275523 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-scripts\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.345001 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.346485 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.348391 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.360734 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.383516 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.385356 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.415583 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.415776 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-config-data\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.415888 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-scripts\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.416030 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd4zs\" (UniqueName: \"kubernetes.io/projected/89eeb481-64d0-4b5a-b470-8edd0091c19b-kube-api-access-nd4zs\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.417987 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.421386 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.423172 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-config-data\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.428468 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-scripts\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.437064 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.471074 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd4zs\" (UniqueName: \"kubernetes.io/projected/89eeb481-64d0-4b5a-b470-8edd0091c19b-kube-api-access-nd4zs\") pod \"nova-cell0-cell-mapping-5szpf\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.488569 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.489973 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.496443 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.500037 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518065 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05bf1de9-28ed-4343-93dc-1dfe47682c59-logs\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518119 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518137 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4m86\" (UniqueName: \"kubernetes.io/projected/05bf1de9-28ed-4343-93dc-1dfe47682c59-kube-api-access-b4m86\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518165 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-config-data\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518213 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-config-data\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518231 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjw8l\" (UniqueName: \"kubernetes.io/projected/6108b630-6716-4ba0-ae41-687c19ebf617-kube-api-access-mjw8l\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.518282 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.528996 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.544255 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.546076 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.550895 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.593722 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620408 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620517 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620550 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt8xn\" (UniqueName: \"kubernetes.io/projected/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-kube-api-access-rt8xn\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620602 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05bf1de9-28ed-4343-93dc-1dfe47682c59-logs\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620620 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620643 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620665 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4m86\" (UniqueName: \"kubernetes.io/projected/05bf1de9-28ed-4343-93dc-1dfe47682c59-kube-api-access-b4m86\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620688 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-config-data\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620733 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-config-data\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.620754 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjw8l\" (UniqueName: \"kubernetes.io/projected/6108b630-6716-4ba0-ae41-687c19ebf617-kube-api-access-mjw8l\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.622190 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05bf1de9-28ed-4343-93dc-1dfe47682c59-logs\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.628949 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.629303 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-config-data\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.633537 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.633616 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-jngsh"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.638216 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.644876 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-config-data\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.651765 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjw8l\" (UniqueName: \"kubernetes.io/projected/6108b630-6716-4ba0-ae41-687c19ebf617-kube-api-access-mjw8l\") pod \"nova-scheduler-0\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.660587 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4m86\" (UniqueName: \"kubernetes.io/projected/05bf1de9-28ed-4343-93dc-1dfe47682c59-kube-api-access-b4m86\") pod \"nova-api-0\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.663882 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.664919 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.668845 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-jngsh"] Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722392 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722484 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722530 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-config-data\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722580 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dszj7\" (UniqueName: \"kubernetes.io/projected/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-kube-api-access-dszj7\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722768 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-logs\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722876 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.722914 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt8xn\" (UniqueName: \"kubernetes.io/projected/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-kube-api-access-rt8xn\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.729177 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.734435 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.738286 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt8xn\" (UniqueName: \"kubernetes.io/projected/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-kube-api-access-rt8xn\") pod \"nova-cell1-novncproxy-0\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824728 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-svc\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824775 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824836 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824890 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824916 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-config-data\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824959 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dszj7\" (UniqueName: \"kubernetes.io/projected/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-kube-api-access-dszj7\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.824994 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfp9g\" (UniqueName: \"kubernetes.io/projected/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-kube-api-access-xfp9g\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.825010 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-config\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.825028 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.825068 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-logs\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.825778 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-logs\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.832331 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-config-data\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.832778 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.845194 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dszj7\" (UniqueName: \"kubernetes.io/projected/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-kube-api-access-dszj7\") pod \"nova-metadata-0\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " pod="openstack/nova-metadata-0" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.928623 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-svc\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.928997 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.929072 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.929187 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfp9g\" (UniqueName: \"kubernetes.io/projected/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-kube-api-access-xfp9g\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.929211 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-config\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.929236 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.929880 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-svc\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.929922 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.930391 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.930889 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-config\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.931580 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:08 crc kubenswrapper[4990]: I1203 13:04:08.952630 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfp9g\" (UniqueName: \"kubernetes.io/projected/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-kube-api-access-xfp9g\") pod \"dnsmasq-dns-865f5d856f-jngsh\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:08.995886 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.010096 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.028883 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.087217 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5szpf"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.266866 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:04:09 crc kubenswrapper[4990]: E1203 13:04:09.267418 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:04:09 crc kubenswrapper[4990]: W1203 13:04:09.359101 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6108b630_6716_4ba0_ae41_687c19ebf617.slice/crio-6311fcc031e9b69c5e0179a4433240bfbf4a58e58d2b68bb1229e82272c96ff0 WatchSource:0}: Error finding container 6311fcc031e9b69c5e0179a4433240bfbf4a58e58d2b68bb1229e82272c96ff0: Status 404 returned error can't find the container with id 6311fcc031e9b69c5e0179a4433240bfbf4a58e58d2b68bb1229e82272c96ff0 Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.362166 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.381072 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.444815 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6108b630-6716-4ba0-ae41-687c19ebf617","Type":"ContainerStarted","Data":"6311fcc031e9b69c5e0179a4433240bfbf4a58e58d2b68bb1229e82272c96ff0"} Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.445912 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05bf1de9-28ed-4343-93dc-1dfe47682c59","Type":"ContainerStarted","Data":"dca62a271d620bafd116962a06184db43b9c0e06d8da4e51ca839017571d4bc3"} Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.446811 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5szpf" event={"ID":"89eeb481-64d0-4b5a-b470-8edd0091c19b","Type":"ContainerStarted","Data":"09d963131d1957a363bf13b5a1f6dfa2b3fd0212160391ae8da45ec543e3a656"} Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.566442 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.626507 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h4gl2"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.628225 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.629903 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.631509 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.639483 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h4gl2"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.717042 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:09 crc kubenswrapper[4990]: W1203 13:04:09.723356 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c5f9deb_d3bd_4e63_8bac_c0f5b71191e4.slice/crio-787720e6786c3f1dd7fc03a95f683164efd80bad0a186d2f7c5970f649fd484a WatchSource:0}: Error finding container 787720e6786c3f1dd7fc03a95f683164efd80bad0a186d2f7c5970f649fd484a: Status 404 returned error can't find the container with id 787720e6786c3f1dd7fc03a95f683164efd80bad0a186d2f7c5970f649fd484a Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.735035 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-jngsh"] Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.749925 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-scripts\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.750068 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-config-data\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.750102 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cshd4\" (UniqueName: \"kubernetes.io/projected/45e64cfc-87ef-43b1-9b8a-e76314fc514e-kube-api-access-cshd4\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.750201 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.851510 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-scripts\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.851977 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-config-data\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.852031 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cshd4\" (UniqueName: \"kubernetes.io/projected/45e64cfc-87ef-43b1-9b8a-e76314fc514e-kube-api-access-cshd4\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.852168 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.855864 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.856139 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-config-data\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.856196 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-scripts\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.873559 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cshd4\" (UniqueName: \"kubernetes.io/projected/45e64cfc-87ef-43b1-9b8a-e76314fc514e-kube-api-access-cshd4\") pod \"nova-cell1-conductor-db-sync-h4gl2\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:09 crc kubenswrapper[4990]: I1203 13:04:09.951981 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.436205 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h4gl2"] Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.460942 4990 generic.go:334] "Generic (PLEG): container finished" podID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerID="13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d" exitCode=0 Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.461023 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" event={"ID":"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5","Type":"ContainerDied","Data":"13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d"} Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.461050 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" event={"ID":"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5","Type":"ContainerStarted","Data":"77e6c024bcf9a88bd785e3abef9d9b7285b393b1569289b7a07271def542f508"} Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.464719 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4","Type":"ContainerStarted","Data":"787720e6786c3f1dd7fc03a95f683164efd80bad0a186d2f7c5970f649fd484a"} Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.468201 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5szpf" event={"ID":"89eeb481-64d0-4b5a-b470-8edd0091c19b","Type":"ContainerStarted","Data":"5a3d5d1b41a54f83b496297d853637e6078bebfdd00162e8f69b807a18d7ef6e"} Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.471970 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf","Type":"ContainerStarted","Data":"c025120a9ea30e0e03ce9427f90d826b1e8f9c13268be9e6ef46301273b98d44"} Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.507913 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5szpf" podStartSLOduration=2.5078968919999998 podStartE2EDuration="2.507896892s" podCreationTimestamp="2025-12-03 13:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:10.501189474 +0000 UTC m=+1598.643100703" watchObservedRunningTime="2025-12-03 13:04:10.507896892 +0000 UTC m=+1598.649808121" Dec 03 13:04:10 crc kubenswrapper[4990]: I1203 13:04:10.537004 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 13:04:11 crc kubenswrapper[4990]: I1203 13:04:11.483249 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" event={"ID":"45e64cfc-87ef-43b1-9b8a-e76314fc514e","Type":"ContainerStarted","Data":"1411d5aade6765c83a24c1e746fc729bf664576cdd59194c1d46d016b97c9854"} Dec 03 13:04:12 crc kubenswrapper[4990]: I1203 13:04:12.181929 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:12 crc kubenswrapper[4990]: I1203 13:04:12.231277 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.503788 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6108b630-6716-4ba0-ae41-687c19ebf617","Type":"ContainerStarted","Data":"f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.506030 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" event={"ID":"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5","Type":"ContainerStarted","Data":"e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.506228 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.508635 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05bf1de9-28ed-4343-93dc-1dfe47682c59","Type":"ContainerStarted","Data":"30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.508775 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05bf1de9-28ed-4343-93dc-1dfe47682c59","Type":"ContainerStarted","Data":"a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.510086 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4","Type":"ContainerStarted","Data":"143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.510249 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019" gracePeriod=30 Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.516563 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf","Type":"ContainerStarted","Data":"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.516908 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf","Type":"ContainerStarted","Data":"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.516803 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-metadata" containerID="cri-o://73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f" gracePeriod=30 Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.516600 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-log" containerID="cri-o://a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411" gracePeriod=30 Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.520189 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" event={"ID":"45e64cfc-87ef-43b1-9b8a-e76314fc514e","Type":"ContainerStarted","Data":"67289d189fa31ade806242db89d99050df375597f4f4c62fe97d4235abbed52b"} Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.549136 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.63718673 podStartE2EDuration="5.549100719s" podCreationTimestamp="2025-12-03 13:04:08 +0000 UTC" firstStartedPulling="2025-12-03 13:04:09.571560259 +0000 UTC m=+1597.713471488" lastFinishedPulling="2025-12-03 13:04:12.483474248 +0000 UTC m=+1600.625385477" observedRunningTime="2025-12-03 13:04:13.546409928 +0000 UTC m=+1601.688321157" watchObservedRunningTime="2025-12-03 13:04:13.549100719 +0000 UTC m=+1601.691011948" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.561890 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.4587994650000002 podStartE2EDuration="5.561855555s" podCreationTimestamp="2025-12-03 13:04:08 +0000 UTC" firstStartedPulling="2025-12-03 13:04:09.361387597 +0000 UTC m=+1597.503298836" lastFinishedPulling="2025-12-03 13:04:12.464443697 +0000 UTC m=+1600.606354926" observedRunningTime="2025-12-03 13:04:13.529130343 +0000 UTC m=+1601.671041602" watchObservedRunningTime="2025-12-03 13:04:13.561855555 +0000 UTC m=+1601.703766784" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.571323 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" podStartSLOduration=4.571304675 podStartE2EDuration="4.571304675s" podCreationTimestamp="2025-12-03 13:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:13.568209983 +0000 UTC m=+1601.710121212" watchObservedRunningTime="2025-12-03 13:04:13.571304675 +0000 UTC m=+1601.713215904" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.593213 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.471782668 podStartE2EDuration="5.593183962s" podCreationTimestamp="2025-12-03 13:04:08 +0000 UTC" firstStartedPulling="2025-12-03 13:04:09.355886922 +0000 UTC m=+1597.497798151" lastFinishedPulling="2025-12-03 13:04:12.477288216 +0000 UTC m=+1600.619199445" observedRunningTime="2025-12-03 13:04:13.587737928 +0000 UTC m=+1601.729649167" watchObservedRunningTime="2025-12-03 13:04:13.593183962 +0000 UTC m=+1601.735095191" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.610381 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.874059367 podStartE2EDuration="5.610346944s" podCreationTimestamp="2025-12-03 13:04:08 +0000 UTC" firstStartedPulling="2025-12-03 13:04:09.739775166 +0000 UTC m=+1597.881686395" lastFinishedPulling="2025-12-03 13:04:12.476062743 +0000 UTC m=+1600.617973972" observedRunningTime="2025-12-03 13:04:13.605316271 +0000 UTC m=+1601.747227500" watchObservedRunningTime="2025-12-03 13:04:13.610346944 +0000 UTC m=+1601.752258173" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.645706 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" podStartSLOduration=5.645674126 podStartE2EDuration="5.645674126s" podCreationTimestamp="2025-12-03 13:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:13.621607661 +0000 UTC m=+1601.763518910" watchObservedRunningTime="2025-12-03 13:04:13.645674126 +0000 UTC m=+1601.787585355" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.667705 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 13:04:13 crc kubenswrapper[4990]: I1203 13:04:13.997553 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.010560 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.010606 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.420399 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.533389 4990 generic.go:334] "Generic (PLEG): container finished" podID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerID="73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f" exitCode=0 Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.534437 4990 generic.go:334] "Generic (PLEG): container finished" podID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerID="a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411" exitCode=143 Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.533517 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf","Type":"ContainerDied","Data":"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f"} Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.536169 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf","Type":"ContainerDied","Data":"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411"} Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.536187 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf","Type":"ContainerDied","Data":"c025120a9ea30e0e03ce9427f90d826b1e8f9c13268be9e6ef46301273b98d44"} Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.536220 4990 scope.go:117] "RemoveContainer" containerID="73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.533654 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.555663 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-logs\") pod \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.556106 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-config-data\") pod \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.556780 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dszj7\" (UniqueName: \"kubernetes.io/projected/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-kube-api-access-dszj7\") pod \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.556890 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-combined-ca-bundle\") pod \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\" (UID: \"86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf\") " Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.557589 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-logs" (OuterVolumeSpecName: "logs") pod "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" (UID: "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.559786 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.563008 4990 scope.go:117] "RemoveContainer" containerID="a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.563054 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-kube-api-access-dszj7" (OuterVolumeSpecName: "kube-api-access-dszj7") pod "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" (UID: "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf"). InnerVolumeSpecName "kube-api-access-dszj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.591131 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" (UID: "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.651919 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-config-data" (OuterVolumeSpecName: "config-data") pod "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" (UID: "86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.660900 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.660928 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dszj7\" (UniqueName: \"kubernetes.io/projected/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-kube-api-access-dszj7\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.660937 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.706945 4990 scope.go:117] "RemoveContainer" containerID="73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f" Dec 03 13:04:14 crc kubenswrapper[4990]: E1203 13:04:14.707485 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f\": container with ID starting with 73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f not found: ID does not exist" containerID="73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.707541 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f"} err="failed to get container status \"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f\": rpc error: code = NotFound desc = could not find container \"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f\": container with ID starting with 73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f not found: ID does not exist" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.707576 4990 scope.go:117] "RemoveContainer" containerID="a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411" Dec 03 13:04:14 crc kubenswrapper[4990]: E1203 13:04:14.708114 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411\": container with ID starting with a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411 not found: ID does not exist" containerID="a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.708148 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411"} err="failed to get container status \"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411\": rpc error: code = NotFound desc = could not find container \"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411\": container with ID starting with a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411 not found: ID does not exist" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.708166 4990 scope.go:117] "RemoveContainer" containerID="73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.708430 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f"} err="failed to get container status \"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f\": rpc error: code = NotFound desc = could not find container \"73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f\": container with ID starting with 73f69dd176f9a204e22843caacc3a6e43b5746439d3bbfb5cedd2cc31985353f not found: ID does not exist" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.708486 4990 scope.go:117] "RemoveContainer" containerID="a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.708966 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411"} err="failed to get container status \"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411\": rpc error: code = NotFound desc = could not find container \"a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411\": container with ID starting with a5ee213bbc84eff002c73df512c5921d30e6501a782ff8ca0b1ad3c5269d9411 not found: ID does not exist" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.869901 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.928520 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.960752 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:14 crc kubenswrapper[4990]: E1203 13:04:14.961198 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-log" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.961217 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-log" Dec 03 13:04:14 crc kubenswrapper[4990]: E1203 13:04:14.961256 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-metadata" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.961262 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-metadata" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.961431 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-log" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.961476 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" containerName="nova-metadata-metadata" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.962404 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.965273 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.965642 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 13:04:14 crc kubenswrapper[4990]: I1203 13:04:14.975848 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.068919 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2264ec51-4abb-4fe9-8e70-9085e3302d51-logs\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.069248 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-config-data\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.069274 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.069443 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v46nv\" (UniqueName: \"kubernetes.io/projected/2264ec51-4abb-4fe9-8e70-9085e3302d51-kube-api-access-v46nv\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.069527 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.171207 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2264ec51-4abb-4fe9-8e70-9085e3302d51-logs\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.171347 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-config-data\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.171737 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2264ec51-4abb-4fe9-8e70-9085e3302d51-logs\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.172316 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.172437 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v46nv\" (UniqueName: \"kubernetes.io/projected/2264ec51-4abb-4fe9-8e70-9085e3302d51-kube-api-access-v46nv\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.172838 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.176576 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-config-data\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.177191 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.183787 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.198126 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v46nv\" (UniqueName: \"kubernetes.io/projected/2264ec51-4abb-4fe9-8e70-9085e3302d51-kube-api-access-v46nv\") pod \"nova-metadata-0\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.226927 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.227161 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9896307b-5428-4f09-a960-1001d654e2be" containerName="kube-state-metrics" containerID="cri-o://b661ba261d228bce4d30aea856e638492446a4cb279683d720a5805c53459265" gracePeriod=30 Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.287594 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.577831 4990 generic.go:334] "Generic (PLEG): container finished" podID="9896307b-5428-4f09-a960-1001d654e2be" containerID="b661ba261d228bce4d30aea856e638492446a4cb279683d720a5805c53459265" exitCode=2 Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.578344 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9896307b-5428-4f09-a960-1001d654e2be","Type":"ContainerDied","Data":"b661ba261d228bce4d30aea856e638492446a4cb279683d720a5805c53459265"} Dec 03 13:04:15 crc kubenswrapper[4990]: I1203 13:04:15.908993 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.141585 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.200206 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnxn9\" (UniqueName: \"kubernetes.io/projected/9896307b-5428-4f09-a960-1001d654e2be-kube-api-access-pnxn9\") pod \"9896307b-5428-4f09-a960-1001d654e2be\" (UID: \"9896307b-5428-4f09-a960-1001d654e2be\") " Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.205872 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9896307b-5428-4f09-a960-1001d654e2be-kube-api-access-pnxn9" (OuterVolumeSpecName: "kube-api-access-pnxn9") pod "9896307b-5428-4f09-a960-1001d654e2be" (UID: "9896307b-5428-4f09-a960-1001d654e2be"). InnerVolumeSpecName "kube-api-access-pnxn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.281727 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf" path="/var/lib/kubelet/pods/86cb7d48-fdb8-4a2d-9de3-afbd43bba3cf/volumes" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.301902 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnxn9\" (UniqueName: \"kubernetes.io/projected/9896307b-5428-4f09-a960-1001d654e2be-kube-api-access-pnxn9\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.589252 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2264ec51-4abb-4fe9-8e70-9085e3302d51","Type":"ContainerStarted","Data":"b3ec61d69eb3d8156cfae75a03fc84fdbdcc8fe6b5c65d5d7eabaea1269df5b3"} Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.592864 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9896307b-5428-4f09-a960-1001d654e2be","Type":"ContainerDied","Data":"5718107ed91601b263fe6a6e9f2d8713c15e4eb710477f9164298f5371602d98"} Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.592932 4990 scope.go:117] "RemoveContainer" containerID="b661ba261d228bce4d30aea856e638492446a4cb279683d720a5805c53459265" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.593008 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.630356 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.653616 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.671753 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:04:16 crc kubenswrapper[4990]: E1203 13:04:16.672192 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9896307b-5428-4f09-a960-1001d654e2be" containerName="kube-state-metrics" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.672212 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9896307b-5428-4f09-a960-1001d654e2be" containerName="kube-state-metrics" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.672386 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9896307b-5428-4f09-a960-1001d654e2be" containerName="kube-state-metrics" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.673031 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.681813 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.682737 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.688475 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.811508 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxsmc\" (UniqueName: \"kubernetes.io/projected/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-api-access-sxsmc\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.811732 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.811809 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.812088 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.914035 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.914082 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.914169 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.914221 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxsmc\" (UniqueName: \"kubernetes.io/projected/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-api-access-sxsmc\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.918797 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.919148 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.919253 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:16 crc kubenswrapper[4990]: I1203 13:04:16.934772 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxsmc\" (UniqueName: \"kubernetes.io/projected/b60c63f3-9d80-4660-9bc9-4afef4bd7cd6-kube-api-access-sxsmc\") pod \"kube-state-metrics-0\" (UID: \"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6\") " pod="openstack/kube-state-metrics-0" Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.044815 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.532811 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 13:04:17 crc kubenswrapper[4990]: W1203 13:04:17.541090 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb60c63f3_9d80_4660_9bc9_4afef4bd7cd6.slice/crio-3322cc67e471f970f74f1aae82080718dc9de9fe4955770f98669ccd72c10e0b WatchSource:0}: Error finding container 3322cc67e471f970f74f1aae82080718dc9de9fe4955770f98669ccd72c10e0b: Status 404 returned error can't find the container with id 3322cc67e471f970f74f1aae82080718dc9de9fe4955770f98669ccd72c10e0b Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.602131 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2264ec51-4abb-4fe9-8e70-9085e3302d51","Type":"ContainerStarted","Data":"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b"} Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.602170 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2264ec51-4abb-4fe9-8e70-9085e3302d51","Type":"ContainerStarted","Data":"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74"} Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.604293 4990 generic.go:334] "Generic (PLEG): container finished" podID="89eeb481-64d0-4b5a-b470-8edd0091c19b" containerID="5a3d5d1b41a54f83b496297d853637e6078bebfdd00162e8f69b807a18d7ef6e" exitCode=0 Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.604373 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5szpf" event={"ID":"89eeb481-64d0-4b5a-b470-8edd0091c19b","Type":"ContainerDied","Data":"5a3d5d1b41a54f83b496297d853637e6078bebfdd00162e8f69b807a18d7ef6e"} Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.605668 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6","Type":"ContainerStarted","Data":"3322cc67e471f970f74f1aae82080718dc9de9fe4955770f98669ccd72c10e0b"} Dec 03 13:04:17 crc kubenswrapper[4990]: I1203 13:04:17.624472 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.624434707 podStartE2EDuration="3.624434707s" podCreationTimestamp="2025-12-03 13:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:17.620553324 +0000 UTC m=+1605.762464573" watchObservedRunningTime="2025-12-03 13:04:17.624434707 +0000 UTC m=+1605.766345936" Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.084893 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.085327 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-central-agent" containerID="cri-o://e809a89e9f94386ae715b68d217fff48ea377ef7a54f73a7c5f830b24faeb48f" gracePeriod=30 Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.085394 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="proxy-httpd" containerID="cri-o://6076acc03eb36ee3c3e352f49f74c6f47311211cc7734e52d0ec1e11bd96629c" gracePeriod=30 Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.085560 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-notification-agent" containerID="cri-o://6d04fa05fe827e2b4dd9e63743d094d0f5a21a64ae5405ef20dd30510790e2f5" gracePeriod=30 Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.085468 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="sg-core" containerID="cri-o://c6c94129175c04aeb0851b0e62bbeaf17cac646b9290dc8c11546d2620988ee1" gracePeriod=30 Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.276737 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9896307b-5428-4f09-a960-1001d654e2be" path="/var/lib/kubelet/pods/9896307b-5428-4f09-a960-1001d654e2be/volumes" Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.616147 4990 generic.go:334] "Generic (PLEG): container finished" podID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerID="6076acc03eb36ee3c3e352f49f74c6f47311211cc7734e52d0ec1e11bd96629c" exitCode=0 Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.616531 4990 generic.go:334] "Generic (PLEG): container finished" podID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerID="c6c94129175c04aeb0851b0e62bbeaf17cac646b9290dc8c11546d2620988ee1" exitCode=2 Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.616221 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerDied","Data":"6076acc03eb36ee3c3e352f49f74c6f47311211cc7734e52d0ec1e11bd96629c"} Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.616573 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerDied","Data":"c6c94129175c04aeb0851b0e62bbeaf17cac646b9290dc8c11546d2620988ee1"} Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.665077 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.665118 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.665489 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.722806 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 13:04:18 crc kubenswrapper[4990]: I1203 13:04:18.991838 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.030625 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.091346 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qppz4"] Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.091588 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerName="dnsmasq-dns" containerID="cri-o://2caad78d97f33a92bfeebdb8aece85f861e9be6a5ce3f7da82c8c7bfb1928ca7" gracePeriod=10 Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.116380 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-config-data\") pod \"89eeb481-64d0-4b5a-b470-8edd0091c19b\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.116487 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-combined-ca-bundle\") pod \"89eeb481-64d0-4b5a-b470-8edd0091c19b\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.116630 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd4zs\" (UniqueName: \"kubernetes.io/projected/89eeb481-64d0-4b5a-b470-8edd0091c19b-kube-api-access-nd4zs\") pod \"89eeb481-64d0-4b5a-b470-8edd0091c19b\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.116798 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-scripts\") pod \"89eeb481-64d0-4b5a-b470-8edd0091c19b\" (UID: \"89eeb481-64d0-4b5a-b470-8edd0091c19b\") " Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.128608 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89eeb481-64d0-4b5a-b470-8edd0091c19b-kube-api-access-nd4zs" (OuterVolumeSpecName: "kube-api-access-nd4zs") pod "89eeb481-64d0-4b5a-b470-8edd0091c19b" (UID: "89eeb481-64d0-4b5a-b470-8edd0091c19b"). InnerVolumeSpecName "kube-api-access-nd4zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.131574 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-scripts" (OuterVolumeSpecName: "scripts") pod "89eeb481-64d0-4b5a-b470-8edd0091c19b" (UID: "89eeb481-64d0-4b5a-b470-8edd0091c19b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.175542 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89eeb481-64d0-4b5a-b470-8edd0091c19b" (UID: "89eeb481-64d0-4b5a-b470-8edd0091c19b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.181204 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-config-data" (OuterVolumeSpecName: "config-data") pod "89eeb481-64d0-4b5a-b470-8edd0091c19b" (UID: "89eeb481-64d0-4b5a-b470-8edd0091c19b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.220244 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd4zs\" (UniqueName: \"kubernetes.io/projected/89eeb481-64d0-4b5a-b470-8edd0091c19b-kube-api-access-nd4zs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.220290 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.220302 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.220312 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89eeb481-64d0-4b5a-b470-8edd0091c19b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.637467 4990 generic.go:334] "Generic (PLEG): container finished" podID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerID="2caad78d97f33a92bfeebdb8aece85f861e9be6a5ce3f7da82c8c7bfb1928ca7" exitCode=0 Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.638321 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" event={"ID":"b02bd168-995c-41d6-b835-a8e4ce46b626","Type":"ContainerDied","Data":"2caad78d97f33a92bfeebdb8aece85f861e9be6a5ce3f7da82c8c7bfb1928ca7"} Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.661520 4990 generic.go:334] "Generic (PLEG): container finished" podID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerID="6d04fa05fe827e2b4dd9e63743d094d0f5a21a64ae5405ef20dd30510790e2f5" exitCode=0 Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.661553 4990 generic.go:334] "Generic (PLEG): container finished" podID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerID="e809a89e9f94386ae715b68d217fff48ea377ef7a54f73a7c5f830b24faeb48f" exitCode=0 Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.661618 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerDied","Data":"6d04fa05fe827e2b4dd9e63743d094d0f5a21a64ae5405ef20dd30510790e2f5"} Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.661646 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerDied","Data":"e809a89e9f94386ae715b68d217fff48ea377ef7a54f73a7c5f830b24faeb48f"} Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.670945 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5szpf" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.671605 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5szpf" event={"ID":"89eeb481-64d0-4b5a-b470-8edd0091c19b","Type":"ContainerDied","Data":"09d963131d1957a363bf13b5a1f6dfa2b3fd0212160391ae8da45ec543e3a656"} Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.671707 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09d963131d1957a363bf13b5a1f6dfa2b3fd0212160391ae8da45ec543e3a656" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.722565 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.747674 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.747927 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.779849 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.780058 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-log" containerID="cri-o://a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a" gracePeriod=30 Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.780496 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-api" containerID="cri-o://30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f" gracePeriod=30 Dec 03 13:04:19 crc kubenswrapper[4990]: E1203 13:04:19.804221 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17e5d538_2a29_4f9d_82b3_ba51790e792a.slice/crio-conmon-6d04fa05fe827e2b4dd9e63743d094d0f5a21a64ae5405ef20dd30510790e2f5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89eeb481_64d0_4b5a_b470_8edd0091c19b.slice\": RecentStats: unable to find data in memory cache]" Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.813662 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.836892 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.852796 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-log" containerID="cri-o://98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74" gracePeriod=30 Dec 03 13:04:19 crc kubenswrapper[4990]: I1203 13:04:19.852989 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-metadata" containerID="cri-o://0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b" gracePeriod=30 Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.090056 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.136307 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-nb\") pod \"b02bd168-995c-41d6-b835-a8e4ce46b626\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.136391 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-svc\") pod \"b02bd168-995c-41d6-b835-a8e4ce46b626\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.136421 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-config\") pod \"b02bd168-995c-41d6-b835-a8e4ce46b626\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.136507 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-sb\") pod \"b02bd168-995c-41d6-b835-a8e4ce46b626\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.136572 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-swift-storage-0\") pod \"b02bd168-995c-41d6-b835-a8e4ce46b626\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.136595 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g25cl\" (UniqueName: \"kubernetes.io/projected/b02bd168-995c-41d6-b835-a8e4ce46b626-kube-api-access-g25cl\") pod \"b02bd168-995c-41d6-b835-a8e4ce46b626\" (UID: \"b02bd168-995c-41d6-b835-a8e4ce46b626\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.143096 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b02bd168-995c-41d6-b835-a8e4ce46b626-kube-api-access-g25cl" (OuterVolumeSpecName: "kube-api-access-g25cl") pod "b02bd168-995c-41d6-b835-a8e4ce46b626" (UID: "b02bd168-995c-41d6-b835-a8e4ce46b626"). InnerVolumeSpecName "kube-api-access-g25cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.213418 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-config" (OuterVolumeSpecName: "config") pod "b02bd168-995c-41d6-b835-a8e4ce46b626" (UID: "b02bd168-995c-41d6-b835-a8e4ce46b626"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.240663 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.240695 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g25cl\" (UniqueName: \"kubernetes.io/projected/b02bd168-995c-41d6-b835-a8e4ce46b626-kube-api-access-g25cl\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.247098 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b02bd168-995c-41d6-b835-a8e4ce46b626" (UID: "b02bd168-995c-41d6-b835-a8e4ce46b626"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.257330 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b02bd168-995c-41d6-b835-a8e4ce46b626" (UID: "b02bd168-995c-41d6-b835-a8e4ce46b626"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.263151 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b02bd168-995c-41d6-b835-a8e4ce46b626" (UID: "b02bd168-995c-41d6-b835-a8e4ce46b626"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.264961 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.265280 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.278940 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b02bd168-995c-41d6-b835-a8e4ce46b626" (UID: "b02bd168-995c-41d6-b835-a8e4ce46b626"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.300988 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.301023 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.362329 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.367781 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.367814 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.367837 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b02bd168-995c-41d6-b835-a8e4ce46b626-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.474420 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.495705 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570389 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-log-httpd\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570705 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-config-data\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570749 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-combined-ca-bundle\") pod \"2264ec51-4abb-4fe9-8e70-9085e3302d51\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570797 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-combined-ca-bundle\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570836 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570877 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2264ec51-4abb-4fe9-8e70-9085e3302d51-logs\") pod \"2264ec51-4abb-4fe9-8e70-9085e3302d51\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570906 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v46nv\" (UniqueName: \"kubernetes.io/projected/2264ec51-4abb-4fe9-8e70-9085e3302d51-kube-api-access-v46nv\") pod \"2264ec51-4abb-4fe9-8e70-9085e3302d51\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570931 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-597lq\" (UniqueName: \"kubernetes.io/projected/17e5d538-2a29-4f9d-82b3-ba51790e792a-kube-api-access-597lq\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570964 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-nova-metadata-tls-certs\") pod \"2264ec51-4abb-4fe9-8e70-9085e3302d51\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.570994 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-sg-core-conf-yaml\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.571025 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-config-data\") pod \"2264ec51-4abb-4fe9-8e70-9085e3302d51\" (UID: \"2264ec51-4abb-4fe9-8e70-9085e3302d51\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.571088 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-run-httpd\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.571104 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-scripts\") pod \"17e5d538-2a29-4f9d-82b3-ba51790e792a\" (UID: \"17e5d538-2a29-4f9d-82b3-ba51790e792a\") " Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.571440 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.574708 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2264ec51-4abb-4fe9-8e70-9085e3302d51-logs" (OuterVolumeSpecName: "logs") pod "2264ec51-4abb-4fe9-8e70-9085e3302d51" (UID: "2264ec51-4abb-4fe9-8e70-9085e3302d51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.575558 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-scripts" (OuterVolumeSpecName: "scripts") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.575912 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.579911 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e5d538-2a29-4f9d-82b3-ba51790e792a-kube-api-access-597lq" (OuterVolumeSpecName: "kube-api-access-597lq") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "kube-api-access-597lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.583731 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2264ec51-4abb-4fe9-8e70-9085e3302d51-kube-api-access-v46nv" (OuterVolumeSpecName: "kube-api-access-v46nv") pod "2264ec51-4abb-4fe9-8e70-9085e3302d51" (UID: "2264ec51-4abb-4fe9-8e70-9085e3302d51"). InnerVolumeSpecName "kube-api-access-v46nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.604295 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2264ec51-4abb-4fe9-8e70-9085e3302d51" (UID: "2264ec51-4abb-4fe9-8e70-9085e3302d51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.617840 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.654722 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-config-data" (OuterVolumeSpecName: "config-data") pod "2264ec51-4abb-4fe9-8e70-9085e3302d51" (UID: "2264ec51-4abb-4fe9-8e70-9085e3302d51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.669638 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672872 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672906 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2264ec51-4abb-4fe9-8e70-9085e3302d51-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672922 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v46nv\" (UniqueName: \"kubernetes.io/projected/2264ec51-4abb-4fe9-8e70-9085e3302d51-kube-api-access-v46nv\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672937 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-597lq\" (UniqueName: \"kubernetes.io/projected/17e5d538-2a29-4f9d-82b3-ba51790e792a-kube-api-access-597lq\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672949 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672961 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672973 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17e5d538-2a29-4f9d-82b3-ba51790e792a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672984 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.672994 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.692721 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2264ec51-4abb-4fe9-8e70-9085e3302d51" (UID: "2264ec51-4abb-4fe9-8e70-9085e3302d51"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.721714 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17e5d538-2a29-4f9d-82b3-ba51790e792a","Type":"ContainerDied","Data":"85bfdeb2105685d680ea7e39cd4ce8d3f9736227fa023fa34fd41572145017ed"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.721794 4990 scope.go:117] "RemoveContainer" containerID="6076acc03eb36ee3c3e352f49f74c6f47311211cc7734e52d0ec1e11bd96629c" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.722028 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.753412 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-config-data" (OuterVolumeSpecName: "config-data") pod "17e5d538-2a29-4f9d-82b3-ba51790e792a" (UID: "17e5d538-2a29-4f9d-82b3-ba51790e792a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.766125 4990 generic.go:334] "Generic (PLEG): container finished" podID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerID="a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a" exitCode=143 Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.766186 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05bf1de9-28ed-4343-93dc-1dfe47682c59","Type":"ContainerDied","Data":"a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.767538 4990 generic.go:334] "Generic (PLEG): container finished" podID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerID="0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b" exitCode=0 Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.767556 4990 generic.go:334] "Generic (PLEG): container finished" podID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerID="98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74" exitCode=143 Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.767583 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2264ec51-4abb-4fe9-8e70-9085e3302d51","Type":"ContainerDied","Data":"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.767600 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2264ec51-4abb-4fe9-8e70-9085e3302d51","Type":"ContainerDied","Data":"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.767610 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2264ec51-4abb-4fe9-8e70-9085e3302d51","Type":"ContainerDied","Data":"b3ec61d69eb3d8156cfae75a03fc84fdbdcc8fe6b5c65d5d7eabaea1269df5b3"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.767608 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.774713 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17e5d538-2a29-4f9d-82b3-ba51790e792a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.774744 4990 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2264ec51-4abb-4fe9-8e70-9085e3302d51-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.778248 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b60c63f3-9d80-4660-9bc9-4afef4bd7cd6","Type":"ContainerStarted","Data":"218d0014a191d36fe1405f7d730feccaf0f1f8b9c8562c290d5c87801dcc4df2"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.778935 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.791299 4990 scope.go:117] "RemoveContainer" containerID="c6c94129175c04aeb0851b0e62bbeaf17cac646b9290dc8c11546d2620988ee1" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.797883 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.798585 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qppz4" event={"ID":"b02bd168-995c-41d6-b835-a8e4ce46b626","Type":"ContainerDied","Data":"08e9bdc1bc2675b276990d98e37799ddb6f126470e294581ce573d1ee5923b9d"} Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.812197 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.7764746970000003 podStartE2EDuration="4.812179079s" podCreationTimestamp="2025-12-03 13:04:16 +0000 UTC" firstStartedPulling="2025-12-03 13:04:17.543738108 +0000 UTC m=+1605.685649337" lastFinishedPulling="2025-12-03 13:04:19.5794425 +0000 UTC m=+1607.721353719" observedRunningTime="2025-12-03 13:04:20.802882133 +0000 UTC m=+1608.944793362" watchObservedRunningTime="2025-12-03 13:04:20.812179079 +0000 UTC m=+1608.954090298" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.828924 4990 scope.go:117] "RemoveContainer" containerID="6d04fa05fe827e2b4dd9e63743d094d0f5a21a64ae5405ef20dd30510790e2f5" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.846971 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.870634 4990 scope.go:117] "RemoveContainer" containerID="e809a89e9f94386ae715b68d217fff48ea377ef7a54f73a7c5f830b24faeb48f" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.881523 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.899858 4990 scope.go:117] "RemoveContainer" containerID="0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.913568 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qppz4"] Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.931929 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qppz4"] Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.948417 4990 scope.go:117] "RemoveContainer" containerID="98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.948789 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949208 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-central-agent" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949227 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-central-agent" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949240 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerName="dnsmasq-dns" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949246 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerName="dnsmasq-dns" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949265 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerName="init" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949271 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerName="init" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949284 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="sg-core" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949290 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="sg-core" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949299 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-notification-agent" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949305 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-notification-agent" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949318 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="proxy-httpd" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949324 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="proxy-httpd" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949335 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-metadata" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949341 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-metadata" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949351 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89eeb481-64d0-4b5a-b470-8edd0091c19b" containerName="nova-manage" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949357 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="89eeb481-64d0-4b5a-b470-8edd0091c19b" containerName="nova-manage" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.949375 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-log" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949380 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-log" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949661 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-central-agent" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949672 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="sg-core" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949681 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-log" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949690 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="ceilometer-notification-agent" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949701 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="89eeb481-64d0-4b5a-b470-8edd0091c19b" containerName="nova-manage" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949711 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" containerName="proxy-httpd" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949723 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" containerName="dnsmasq-dns" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.949738 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" containerName="nova-metadata-metadata" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.950780 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.956055 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.956311 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.958341 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.978139 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-config-data\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.978232 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d64wd\" (UniqueName: \"kubernetes.io/projected/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-kube-api-access-d64wd\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.978273 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-logs\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.978325 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.978361 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.994075 4990 scope.go:117] "RemoveContainer" containerID="0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.994452 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b\": container with ID starting with 0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b not found: ID does not exist" containerID="0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.994598 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b"} err="failed to get container status \"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b\": rpc error: code = NotFound desc = could not find container \"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b\": container with ID starting with 0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b not found: ID does not exist" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.994636 4990 scope.go:117] "RemoveContainer" containerID="98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74" Dec 03 13:04:20 crc kubenswrapper[4990]: E1203 13:04:20.995193 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74\": container with ID starting with 98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74 not found: ID does not exist" containerID="98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.995254 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74"} err="failed to get container status \"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74\": rpc error: code = NotFound desc = could not find container \"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74\": container with ID starting with 98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74 not found: ID does not exist" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.995273 4990 scope.go:117] "RemoveContainer" containerID="0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.998713 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b"} err="failed to get container status \"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b\": rpc error: code = NotFound desc = could not find container \"0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b\": container with ID starting with 0a3551b44a6822ca388a6ebc73d97b490fdc96143300f57f95cf471cea03c38b not found: ID does not exist" Dec 03 13:04:20 crc kubenswrapper[4990]: I1203 13:04:20.998746 4990 scope.go:117] "RemoveContainer" containerID="98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.002629 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74"} err="failed to get container status \"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74\": rpc error: code = NotFound desc = could not find container \"98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74\": container with ID starting with 98590d19a3a4d42afd48db2cacfff441087d2ccfac5db5b9799f860e0962fb74 not found: ID does not exist" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.002674 4990 scope.go:117] "RemoveContainer" containerID="2caad78d97f33a92bfeebdb8aece85f861e9be6a5ce3f7da82c8c7bfb1928ca7" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.025688 4990 scope.go:117] "RemoveContainer" containerID="4aee085128209fa82325b11a3cea858fffb7e789070f83cc7030c0c5eb721e7f" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.072695 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.080026 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-logs\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.080119 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.080162 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.080285 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-config-data\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.080374 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d64wd\" (UniqueName: \"kubernetes.io/projected/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-kube-api-access-d64wd\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.080535 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-logs\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.087847 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.090544 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-config-data\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.090682 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.102053 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.104740 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d64wd\" (UniqueName: \"kubernetes.io/projected/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-kube-api-access-d64wd\") pod \"nova-metadata-0\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.111020 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.114106 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.121738 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.125524 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.125728 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.131929 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.181740 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-scripts\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.181807 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcbbw\" (UniqueName: \"kubernetes.io/projected/87c13d56-b975-4be0-94f5-efa2e80b4b31-kube-api-access-gcbbw\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.181901 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-run-httpd\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.181924 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.181950 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-config-data\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.181973 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-log-httpd\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.182005 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.182080 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.279349 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284657 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-scripts\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284720 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcbbw\" (UniqueName: \"kubernetes.io/projected/87c13d56-b975-4be0-94f5-efa2e80b4b31-kube-api-access-gcbbw\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284805 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-run-httpd\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284829 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284857 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-config-data\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284881 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-log-httpd\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284908 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.284971 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.286521 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-log-httpd\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.287329 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-run-httpd\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.290270 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-scripts\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.296478 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-config-data\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.299967 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.300510 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.304172 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcbbw\" (UniqueName: \"kubernetes.io/projected/87c13d56-b975-4be0-94f5-efa2e80b4b31-kube-api-access-gcbbw\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.305747 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.448603 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.788771 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.818856 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0780aac2-023d-4f4a-8fef-ccbaf8af60c7","Type":"ContainerStarted","Data":"f610017222edeb5eb8ef3b109e1b2227e621b1457b053c543008297927963ca1"} Dec 03 13:04:21 crc kubenswrapper[4990]: I1203 13:04:21.822682 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6108b630-6716-4ba0-ae41-687c19ebf617" containerName="nova-scheduler-scheduler" containerID="cri-o://f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105" gracePeriod=30 Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.010737 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:22 crc kubenswrapper[4990]: W1203 13:04:22.013863 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87c13d56_b975_4be0_94f5_efa2e80b4b31.slice/crio-cf4dc51eff43cdb7eec0f0dc94c7c7da9011fb7dedca2e1ed57438dbcca477c2 WatchSource:0}: Error finding container cf4dc51eff43cdb7eec0f0dc94c7c7da9011fb7dedca2e1ed57438dbcca477c2: Status 404 returned error can't find the container with id cf4dc51eff43cdb7eec0f0dc94c7c7da9011fb7dedca2e1ed57438dbcca477c2 Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.283651 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e5d538-2a29-4f9d-82b3-ba51790e792a" path="/var/lib/kubelet/pods/17e5d538-2a29-4f9d-82b3-ba51790e792a/volumes" Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.284866 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2264ec51-4abb-4fe9-8e70-9085e3302d51" path="/var/lib/kubelet/pods/2264ec51-4abb-4fe9-8e70-9085e3302d51/volumes" Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.287111 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b02bd168-995c-41d6-b835-a8e4ce46b626" path="/var/lib/kubelet/pods/b02bd168-995c-41d6-b835-a8e4ce46b626/volumes" Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.837550 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0780aac2-023d-4f4a-8fef-ccbaf8af60c7","Type":"ContainerStarted","Data":"bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e"} Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.837849 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0780aac2-023d-4f4a-8fef-ccbaf8af60c7","Type":"ContainerStarted","Data":"df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146"} Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.839016 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerStarted","Data":"cf4dc51eff43cdb7eec0f0dc94c7c7da9011fb7dedca2e1ed57438dbcca477c2"} Dec 03 13:04:22 crc kubenswrapper[4990]: I1203 13:04:22.866263 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.866247455 podStartE2EDuration="2.866247455s" podCreationTimestamp="2025-12-03 13:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:22.860831262 +0000 UTC m=+1611.002742491" watchObservedRunningTime="2025-12-03 13:04:22.866247455 +0000 UTC m=+1611.008158684" Dec 03 13:04:23 crc kubenswrapper[4990]: E1203 13:04:23.667264 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 13:04:23 crc kubenswrapper[4990]: E1203 13:04:23.669572 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 13:04:23 crc kubenswrapper[4990]: E1203 13:04:23.672028 4990 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 13:04:23 crc kubenswrapper[4990]: E1203 13:04:23.672112 4990 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6108b630-6716-4ba0-ae41-687c19ebf617" containerName="nova-scheduler-scheduler" Dec 03 13:04:23 crc kubenswrapper[4990]: I1203 13:04:23.860316 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerStarted","Data":"ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305"} Dec 03 13:04:23 crc kubenswrapper[4990]: I1203 13:04:23.860373 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerStarted","Data":"697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea"} Dec 03 13:04:24 crc kubenswrapper[4990]: I1203 13:04:24.871801 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerStarted","Data":"ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006"} Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.743227 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.788403 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-combined-ca-bundle\") pod \"05bf1de9-28ed-4343-93dc-1dfe47682c59\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.788539 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-config-data\") pod \"05bf1de9-28ed-4343-93dc-1dfe47682c59\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.789369 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4m86\" (UniqueName: \"kubernetes.io/projected/05bf1de9-28ed-4343-93dc-1dfe47682c59-kube-api-access-b4m86\") pod \"05bf1de9-28ed-4343-93dc-1dfe47682c59\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.789437 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05bf1de9-28ed-4343-93dc-1dfe47682c59-logs\") pod \"05bf1de9-28ed-4343-93dc-1dfe47682c59\" (UID: \"05bf1de9-28ed-4343-93dc-1dfe47682c59\") " Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.790088 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05bf1de9-28ed-4343-93dc-1dfe47682c59-logs" (OuterVolumeSpecName: "logs") pod "05bf1de9-28ed-4343-93dc-1dfe47682c59" (UID: "05bf1de9-28ed-4343-93dc-1dfe47682c59"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.796622 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05bf1de9-28ed-4343-93dc-1dfe47682c59-kube-api-access-b4m86" (OuterVolumeSpecName: "kube-api-access-b4m86") pod "05bf1de9-28ed-4343-93dc-1dfe47682c59" (UID: "05bf1de9-28ed-4343-93dc-1dfe47682c59"). InnerVolumeSpecName "kube-api-access-b4m86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.821350 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-config-data" (OuterVolumeSpecName: "config-data") pod "05bf1de9-28ed-4343-93dc-1dfe47682c59" (UID: "05bf1de9-28ed-4343-93dc-1dfe47682c59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.824186 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05bf1de9-28ed-4343-93dc-1dfe47682c59" (UID: "05bf1de9-28ed-4343-93dc-1dfe47682c59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.880859 4990 generic.go:334] "Generic (PLEG): container finished" podID="6108b630-6716-4ba0-ae41-687c19ebf617" containerID="f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105" exitCode=0 Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.880923 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6108b630-6716-4ba0-ae41-687c19ebf617","Type":"ContainerDied","Data":"f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105"} Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.882953 4990 generic.go:334] "Generic (PLEG): container finished" podID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerID="30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f" exitCode=0 Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.883017 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.883042 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05bf1de9-28ed-4343-93dc-1dfe47682c59","Type":"ContainerDied","Data":"30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f"} Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.883477 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05bf1de9-28ed-4343-93dc-1dfe47682c59","Type":"ContainerDied","Data":"dca62a271d620bafd116962a06184db43b9c0e06d8da4e51ca839017571d4bc3"} Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.883497 4990 scope.go:117] "RemoveContainer" containerID="30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.885929 4990 generic.go:334] "Generic (PLEG): container finished" podID="45e64cfc-87ef-43b1-9b8a-e76314fc514e" containerID="67289d189fa31ade806242db89d99050df375597f4f4c62fe97d4235abbed52b" exitCode=0 Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.885993 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" event={"ID":"45e64cfc-87ef-43b1-9b8a-e76314fc514e","Type":"ContainerDied","Data":"67289d189fa31ade806242db89d99050df375597f4f4c62fe97d4235abbed52b"} Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.892197 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4m86\" (UniqueName: \"kubernetes.io/projected/05bf1de9-28ed-4343-93dc-1dfe47682c59-kube-api-access-b4m86\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.892229 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05bf1de9-28ed-4343-93dc-1dfe47682c59-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.892240 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.892249 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05bf1de9-28ed-4343-93dc-1dfe47682c59-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.920735 4990 scope.go:117] "RemoveContainer" containerID="a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.925059 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.942398 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.982670 4990 scope.go:117] "RemoveContainer" containerID="30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f" Dec 03 13:04:25 crc kubenswrapper[4990]: E1203 13:04:25.984021 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f\": container with ID starting with 30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f not found: ID does not exist" containerID="30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.984154 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f"} err="failed to get container status \"30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f\": rpc error: code = NotFound desc = could not find container \"30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f\": container with ID starting with 30e2b87779e03b0f279bc9119707b40ad9fa9b14b01317d31b24e2c28996ff2f not found: ID does not exist" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.984200 4990 scope.go:117] "RemoveContainer" containerID="a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a" Dec 03 13:04:25 crc kubenswrapper[4990]: E1203 13:04:25.985285 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a\": container with ID starting with a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a not found: ID does not exist" containerID="a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.985355 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a"} err="failed to get container status \"a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a\": rpc error: code = NotFound desc = could not find container \"a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a\": container with ID starting with a7a620b3f2ade07df26cb2e621176768de58f785c55228f55efbc754e34ef24a not found: ID does not exist" Dec 03 13:04:25 crc kubenswrapper[4990]: I1203 13:04:25.999215 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:26 crc kubenswrapper[4990]: E1203 13:04:26.001530 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-log" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.001568 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-log" Dec 03 13:04:26 crc kubenswrapper[4990]: E1203 13:04:26.001606 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-api" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.001616 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-api" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.003007 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-log" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.004167 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" containerName="nova-api-api" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.010490 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.017289 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.037196 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.210572 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-config-data\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.210631 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsdfc\" (UniqueName: \"kubernetes.io/projected/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-kube-api-access-lsdfc\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.210715 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.210735 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-logs\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.277963 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05bf1de9-28ed-4343-93dc-1dfe47682c59" path="/var/lib/kubelet/pods/05bf1de9-28ed-4343-93dc-1dfe47682c59/volumes" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.281061 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.281111 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.307537 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.312149 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-config-data\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.312191 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsdfc\" (UniqueName: \"kubernetes.io/projected/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-kube-api-access-lsdfc\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.312273 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.312293 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-logs\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.312759 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-logs\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.315551 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-config-data\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.315899 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.333542 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsdfc\" (UniqueName: \"kubernetes.io/projected/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-kube-api-access-lsdfc\") pod \"nova-api-0\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.336311 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.423275 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-combined-ca-bundle\") pod \"6108b630-6716-4ba0-ae41-687c19ebf617\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.423330 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-config-data\") pod \"6108b630-6716-4ba0-ae41-687c19ebf617\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.423600 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjw8l\" (UniqueName: \"kubernetes.io/projected/6108b630-6716-4ba0-ae41-687c19ebf617-kube-api-access-mjw8l\") pod \"6108b630-6716-4ba0-ae41-687c19ebf617\" (UID: \"6108b630-6716-4ba0-ae41-687c19ebf617\") " Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.434202 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6108b630-6716-4ba0-ae41-687c19ebf617-kube-api-access-mjw8l" (OuterVolumeSpecName: "kube-api-access-mjw8l") pod "6108b630-6716-4ba0-ae41-687c19ebf617" (UID: "6108b630-6716-4ba0-ae41-687c19ebf617"). InnerVolumeSpecName "kube-api-access-mjw8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.453273 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-config-data" (OuterVolumeSpecName: "config-data") pod "6108b630-6716-4ba0-ae41-687c19ebf617" (UID: "6108b630-6716-4ba0-ae41-687c19ebf617"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.463315 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6108b630-6716-4ba0-ae41-687c19ebf617" (UID: "6108b630-6716-4ba0-ae41-687c19ebf617"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.526583 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.526617 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6108b630-6716-4ba0-ae41-687c19ebf617-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.526630 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjw8l\" (UniqueName: \"kubernetes.io/projected/6108b630-6716-4ba0-ae41-687c19ebf617-kube-api-access-mjw8l\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.934511 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerStarted","Data":"f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261"} Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.934729 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.936740 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.974282 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.974587 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6108b630-6716-4ba0-ae41-687c19ebf617","Type":"ContainerDied","Data":"6311fcc031e9b69c5e0179a4433240bfbf4a58e58d2b68bb1229e82272c96ff0"} Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.974658 4990 scope.go:117] "RemoveContainer" containerID="f3c58526f672892f154827e8499cc69f82ae7cce5389aa5aa99c50b6f80c6105" Dec 03 13:04:26 crc kubenswrapper[4990]: I1203 13:04:26.987521 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.703372589 podStartE2EDuration="5.987499312s" podCreationTimestamp="2025-12-03 13:04:21 +0000 UTC" firstStartedPulling="2025-12-03 13:04:22.018738645 +0000 UTC m=+1610.160649874" lastFinishedPulling="2025-12-03 13:04:26.302865368 +0000 UTC m=+1614.444776597" observedRunningTime="2025-12-03 13:04:26.98019885 +0000 UTC m=+1615.122110089" watchObservedRunningTime="2025-12-03 13:04:26.987499312 +0000 UTC m=+1615.129410561" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.096549 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.097514 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.132427 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.155519 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:27 crc kubenswrapper[4990]: E1203 13:04:27.155964 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6108b630-6716-4ba0-ae41-687c19ebf617" containerName="nova-scheduler-scheduler" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.155976 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6108b630-6716-4ba0-ae41-687c19ebf617" containerName="nova-scheduler-scheduler" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.156178 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6108b630-6716-4ba0-ae41-687c19ebf617" containerName="nova-scheduler-scheduler" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.156828 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.163480 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.174405 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.250522 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrqph\" (UniqueName: \"kubernetes.io/projected/32409013-6281-4048-8606-ae246d1b1aed-kube-api-access-wrqph\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.250637 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-config-data\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.250711 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.355702 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-config-data\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.355788 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.355904 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrqph\" (UniqueName: \"kubernetes.io/projected/32409013-6281-4048-8606-ae246d1b1aed-kube-api-access-wrqph\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.363710 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-config-data\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.373518 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.374118 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrqph\" (UniqueName: \"kubernetes.io/projected/32409013-6281-4048-8606-ae246d1b1aed-kube-api-access-wrqph\") pod \"nova-scheduler-0\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.521503 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.525215 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.660160 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-scripts\") pod \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.660228 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-config-data\") pod \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.660254 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cshd4\" (UniqueName: \"kubernetes.io/projected/45e64cfc-87ef-43b1-9b8a-e76314fc514e-kube-api-access-cshd4\") pod \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.660441 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-combined-ca-bundle\") pod \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\" (UID: \"45e64cfc-87ef-43b1-9b8a-e76314fc514e\") " Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.665300 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45e64cfc-87ef-43b1-9b8a-e76314fc514e-kube-api-access-cshd4" (OuterVolumeSpecName: "kube-api-access-cshd4") pod "45e64cfc-87ef-43b1-9b8a-e76314fc514e" (UID: "45e64cfc-87ef-43b1-9b8a-e76314fc514e"). InnerVolumeSpecName "kube-api-access-cshd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.676731 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-scripts" (OuterVolumeSpecName: "scripts") pod "45e64cfc-87ef-43b1-9b8a-e76314fc514e" (UID: "45e64cfc-87ef-43b1-9b8a-e76314fc514e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.699550 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-config-data" (OuterVolumeSpecName: "config-data") pod "45e64cfc-87ef-43b1-9b8a-e76314fc514e" (UID: "45e64cfc-87ef-43b1-9b8a-e76314fc514e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.714578 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45e64cfc-87ef-43b1-9b8a-e76314fc514e" (UID: "45e64cfc-87ef-43b1-9b8a-e76314fc514e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.764121 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.764407 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.764421 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e64cfc-87ef-43b1-9b8a-e76314fc514e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.764432 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cshd4\" (UniqueName: \"kubernetes.io/projected/45e64cfc-87ef-43b1-9b8a-e76314fc514e-kube-api-access-cshd4\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.986625 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" event={"ID":"45e64cfc-87ef-43b1-9b8a-e76314fc514e","Type":"ContainerDied","Data":"1411d5aade6765c83a24c1e746fc729bf664576cdd59194c1d46d016b97c9854"} Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.986674 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1411d5aade6765c83a24c1e746fc729bf664576cdd59194c1d46d016b97c9854" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.986739 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-h4gl2" Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.995418 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce28826a-46d4-4ccb-8f2f-65cd0f035c36","Type":"ContainerStarted","Data":"d640fdf5618d25d7277da5d497e23562a7c88e997a561356044c85aa0c949b78"} Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.995472 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce28826a-46d4-4ccb-8f2f-65cd0f035c36","Type":"ContainerStarted","Data":"c02236eb6c3d6123a966261adcb0298733d380099030b20571666800bee2441e"} Dec 03 13:04:27 crc kubenswrapper[4990]: I1203 13:04:27.995486 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce28826a-46d4-4ccb-8f2f-65cd0f035c36","Type":"ContainerStarted","Data":"bfc16465b787828760188d59779febd349fc0f5561f24907d791086a247345b2"} Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.000569 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 13:04:28 crc kubenswrapper[4990]: E1203 13:04:28.000986 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e64cfc-87ef-43b1-9b8a-e76314fc514e" containerName="nova-cell1-conductor-db-sync" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.001008 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e64cfc-87ef-43b1-9b8a-e76314fc514e" containerName="nova-cell1-conductor-db-sync" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.001269 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e64cfc-87ef-43b1-9b8a-e76314fc514e" containerName="nova-cell1-conductor-db-sync" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.001977 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.006039 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.014852 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.035638 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.035623482 podStartE2EDuration="3.035623482s" podCreationTimestamp="2025-12-03 13:04:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:28.025751892 +0000 UTC m=+1616.167663171" watchObservedRunningTime="2025-12-03 13:04:28.035623482 +0000 UTC m=+1616.177534711" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.070665 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34660cc-b629-4ac1-ab68-0546243d600a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.070753 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34660cc-b629-4ac1-ab68-0546243d600a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.070906 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7jf8\" (UniqueName: \"kubernetes.io/projected/a34660cc-b629-4ac1-ab68-0546243d600a-kube-api-access-x7jf8\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.085896 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:04:28 crc kubenswrapper[4990]: W1203 13:04:28.106749 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32409013_6281_4048_8606_ae246d1b1aed.slice/crio-40d83e7741eae69f4c0f7122c7d19f8db18c60df20ef5a776a31e5fff2316a2f WatchSource:0}: Error finding container 40d83e7741eae69f4c0f7122c7d19f8db18c60df20ef5a776a31e5fff2316a2f: Status 404 returned error can't find the container with id 40d83e7741eae69f4c0f7122c7d19f8db18c60df20ef5a776a31e5fff2316a2f Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.172993 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34660cc-b629-4ac1-ab68-0546243d600a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.173179 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7jf8\" (UniqueName: \"kubernetes.io/projected/a34660cc-b629-4ac1-ab68-0546243d600a-kube-api-access-x7jf8\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.173240 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34660cc-b629-4ac1-ab68-0546243d600a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.177746 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34660cc-b629-4ac1-ab68-0546243d600a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.179885 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34660cc-b629-4ac1-ab68-0546243d600a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.192667 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7jf8\" (UniqueName: \"kubernetes.io/projected/a34660cc-b629-4ac1-ab68-0546243d600a-kube-api-access-x7jf8\") pod \"nova-cell1-conductor-0\" (UID: \"a34660cc-b629-4ac1-ab68-0546243d600a\") " pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.293061 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6108b630-6716-4ba0-ae41-687c19ebf617" path="/var/lib/kubelet/pods/6108b630-6716-4ba0-ae41-687c19ebf617/volumes" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.323038 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:28 crc kubenswrapper[4990]: I1203 13:04:28.787110 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 13:04:29 crc kubenswrapper[4990]: I1203 13:04:29.007286 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a34660cc-b629-4ac1-ab68-0546243d600a","Type":"ContainerStarted","Data":"48d18e94846934108a22f50d41e1e970b662152d4c75fe1278adca2231aa467c"} Dec 03 13:04:29 crc kubenswrapper[4990]: I1203 13:04:29.009350 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32409013-6281-4048-8606-ae246d1b1aed","Type":"ContainerStarted","Data":"ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf"} Dec 03 13:04:29 crc kubenswrapper[4990]: I1203 13:04:29.009390 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32409013-6281-4048-8606-ae246d1b1aed","Type":"ContainerStarted","Data":"40d83e7741eae69f4c0f7122c7d19f8db18c60df20ef5a776a31e5fff2316a2f"} Dec 03 13:04:29 crc kubenswrapper[4990]: I1203 13:04:29.032961 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.032944521 podStartE2EDuration="2.032944521s" podCreationTimestamp="2025-12-03 13:04:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:29.028138765 +0000 UTC m=+1617.170049994" watchObservedRunningTime="2025-12-03 13:04:29.032944521 +0000 UTC m=+1617.174855750" Dec 03 13:04:30 crc kubenswrapper[4990]: I1203 13:04:30.023143 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a34660cc-b629-4ac1-ab68-0546243d600a","Type":"ContainerStarted","Data":"6195a76ed6e821cc3d3fcbcab33df84048fd665f65dcc1e45df1027ff9fc3795"} Dec 03 13:04:30 crc kubenswrapper[4990]: I1203 13:04:30.023569 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:30 crc kubenswrapper[4990]: I1203 13:04:30.044517 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.044489766 podStartE2EDuration="3.044489766s" podCreationTimestamp="2025-12-03 13:04:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:30.037582514 +0000 UTC m=+1618.179493753" watchObservedRunningTime="2025-12-03 13:04:30.044489766 +0000 UTC m=+1618.186400995" Dec 03 13:04:31 crc kubenswrapper[4990]: I1203 13:04:31.280662 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:04:31 crc kubenswrapper[4990]: I1203 13:04:31.281142 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:04:32 crc kubenswrapper[4990]: I1203 13:04:32.326814 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:04:32 crc kubenswrapper[4990]: I1203 13:04:32.326837 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:04:32 crc kubenswrapper[4990]: I1203 13:04:32.526410 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 13:04:35 crc kubenswrapper[4990]: I1203 13:04:35.264668 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:04:35 crc kubenswrapper[4990]: E1203 13:04:35.265250 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:04:36 crc kubenswrapper[4990]: I1203 13:04:36.337807 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:04:36 crc kubenswrapper[4990]: I1203 13:04:36.338127 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:04:37 crc kubenswrapper[4990]: I1203 13:04:37.420699 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:04:37 crc kubenswrapper[4990]: I1203 13:04:37.420697 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 13:04:37 crc kubenswrapper[4990]: I1203 13:04:37.526040 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 13:04:37 crc kubenswrapper[4990]: I1203 13:04:37.553634 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 13:04:38 crc kubenswrapper[4990]: I1203 13:04:38.117662 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 13:04:38 crc kubenswrapper[4990]: I1203 13:04:38.349427 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 13:04:41 crc kubenswrapper[4990]: I1203 13:04:41.288154 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 13:04:41 crc kubenswrapper[4990]: I1203 13:04:41.291257 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 13:04:41 crc kubenswrapper[4990]: I1203 13:04:41.294577 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 13:04:42 crc kubenswrapper[4990]: I1203 13:04:42.125672 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 13:04:43 crc kubenswrapper[4990]: I1203 13:04:43.979766 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.077728 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-config-data\") pod \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.077910 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt8xn\" (UniqueName: \"kubernetes.io/projected/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-kube-api-access-rt8xn\") pod \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.078020 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-combined-ca-bundle\") pod \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\" (UID: \"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4\") " Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.086190 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-kube-api-access-rt8xn" (OuterVolumeSpecName: "kube-api-access-rt8xn") pod "6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" (UID: "6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4"). InnerVolumeSpecName "kube-api-access-rt8xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.114234 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-config-data" (OuterVolumeSpecName: "config-data") pod "6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" (UID: "6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.123313 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" (UID: "6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.139433 4990 generic.go:334] "Generic (PLEG): container finished" podID="6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" containerID="143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019" exitCode=137 Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.140396 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.140493 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4","Type":"ContainerDied","Data":"143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019"} Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.140534 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4","Type":"ContainerDied","Data":"787720e6786c3f1dd7fc03a95f683164efd80bad0a186d2f7c5970f649fd484a"} Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.140555 4990 scope.go:117] "RemoveContainer" containerID="143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.179800 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.179837 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt8xn\" (UniqueName: \"kubernetes.io/projected/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-kube-api-access-rt8xn\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.179850 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.209241 4990 scope.go:117] "RemoveContainer" containerID="143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019" Dec 03 13:04:44 crc kubenswrapper[4990]: E1203 13:04:44.210022 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019\": container with ID starting with 143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019 not found: ID does not exist" containerID="143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.210062 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019"} err="failed to get container status \"143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019\": rpc error: code = NotFound desc = could not find container \"143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019\": container with ID starting with 143e5e0f0c98918433625693466f329b55dd974e154d8c95f098fcd2baa3b019 not found: ID does not exist" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.223613 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.232949 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.244099 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:44 crc kubenswrapper[4990]: E1203 13:04:44.244690 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.244709 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.244890 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.245588 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.254893 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.254907 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.255094 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.259479 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.280806 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42fq2\" (UniqueName: \"kubernetes.io/projected/37a5c5c8-e837-456a-94ac-bd2d436b7e44-kube-api-access-42fq2\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.280900 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.280917 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.280952 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.281008 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.286125 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4" path="/var/lib/kubelet/pods/6c5f9deb-d3bd-4e63-8bac-c0f5b71191e4/volumes" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.382428 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.382521 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42fq2\" (UniqueName: \"kubernetes.io/projected/37a5c5c8-e837-456a-94ac-bd2d436b7e44-kube-api-access-42fq2\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.382614 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.382631 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.382680 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.387169 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.387523 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.387777 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.388706 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a5c5c8-e837-456a-94ac-bd2d436b7e44-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.399713 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42fq2\" (UniqueName: \"kubernetes.io/projected/37a5c5c8-e837-456a-94ac-bd2d436b7e44-kube-api-access-42fq2\") pod \"nova-cell1-novncproxy-0\" (UID: \"37a5c5c8-e837-456a-94ac-bd2d436b7e44\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:44 crc kubenswrapper[4990]: I1203 13:04:44.567283 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.034094 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 13:04:45 crc kubenswrapper[4990]: W1203 13:04:45.047749 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37a5c5c8_e837_456a_94ac_bd2d436b7e44.slice/crio-36367d23a3fb3d5654bc76f9619e52ec8722741933be9eecdc1704908c9c67ea WatchSource:0}: Error finding container 36367d23a3fb3d5654bc76f9619e52ec8722741933be9eecdc1704908c9c67ea: Status 404 returned error can't find the container with id 36367d23a3fb3d5654bc76f9619e52ec8722741933be9eecdc1704908c9c67ea Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.077412 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pg5vd"] Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.082125 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.095235 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pg5vd"] Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.096932 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-utilities\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.097017 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26rjq\" (UniqueName: \"kubernetes.io/projected/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-kube-api-access-26rjq\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.097058 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-catalog-content\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.201012 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-utilities\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.201154 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26rjq\" (UniqueName: \"kubernetes.io/projected/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-kube-api-access-26rjq\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.201195 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-catalog-content\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.202170 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-catalog-content\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.202662 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-utilities\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.211327 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"37a5c5c8-e837-456a-94ac-bd2d436b7e44","Type":"ContainerStarted","Data":"36367d23a3fb3d5654bc76f9619e52ec8722741933be9eecdc1704908c9c67ea"} Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.230475 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26rjq\" (UniqueName: \"kubernetes.io/projected/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-kube-api-access-26rjq\") pod \"community-operators-pg5vd\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:45 crc kubenswrapper[4990]: I1203 13:04:45.518423 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:46 crc kubenswrapper[4990]: W1203 13:04:46.106631 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc684cd85_ef9f_448b_bc2d_daaf12bf27f4.slice/crio-90a05cbc2a9812dd3f7c8d0694131f4b0e8112374f6d4b2652dbe4cbd4e8d312 WatchSource:0}: Error finding container 90a05cbc2a9812dd3f7c8d0694131f4b0e8112374f6d4b2652dbe4cbd4e8d312: Status 404 returned error can't find the container with id 90a05cbc2a9812dd3f7c8d0694131f4b0e8112374f6d4b2652dbe4cbd4e8d312 Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.110170 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pg5vd"] Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.236653 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"37a5c5c8-e837-456a-94ac-bd2d436b7e44","Type":"ContainerStarted","Data":"3cb3ec47d02742e2d4991b4215e35d480245dcdb8c85e329f8e56740eceeaa2a"} Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.240257 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerStarted","Data":"90a05cbc2a9812dd3f7c8d0694131f4b0e8112374f6d4b2652dbe4cbd4e8d312"} Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.341235 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.341851 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.344874 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.346907 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:04:46 crc kubenswrapper[4990]: I1203 13:04:46.369571 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.369549397 podStartE2EDuration="2.369549397s" podCreationTimestamp="2025-12-03 13:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:46.26398217 +0000 UTC m=+1634.405893419" watchObservedRunningTime="2025-12-03 13:04:46.369549397 +0000 UTC m=+1634.511460646" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.250709 4990 generic.go:334] "Generic (PLEG): container finished" podID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerID="3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11" exitCode=0 Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.250823 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerDied","Data":"3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11"} Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.251102 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.254801 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.462710 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-b24tp"] Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.464302 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.487232 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-b24tp"] Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.549323 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.549670 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.549761 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.549931 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-config\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.550163 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qcg4\" (UniqueName: \"kubernetes.io/projected/32c217f2-3a1b-47f5-91d9-b6634935c9f0-kube-api-access-7qcg4\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.550215 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.652444 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.652565 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.652602 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.652703 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.652734 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-config\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.652804 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qcg4\" (UniqueName: \"kubernetes.io/projected/32c217f2-3a1b-47f5-91d9-b6634935c9f0-kube-api-access-7qcg4\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.653287 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.653557 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.653937 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.654035 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.654091 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-config\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.700956 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qcg4\" (UniqueName: \"kubernetes.io/projected/32c217f2-3a1b-47f5-91d9-b6634935c9f0-kube-api-access-7qcg4\") pod \"dnsmasq-dns-5c7b6c5df9-b24tp\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:47 crc kubenswrapper[4990]: I1203 13:04:47.806302 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:48 crc kubenswrapper[4990]: I1203 13:04:48.263428 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerStarted","Data":"3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2"} Dec 03 13:04:48 crc kubenswrapper[4990]: I1203 13:04:48.356854 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-b24tp"] Dec 03 13:04:49 crc kubenswrapper[4990]: I1203 13:04:49.281532 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:04:49 crc kubenswrapper[4990]: E1203 13:04:49.282108 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:04:49 crc kubenswrapper[4990]: I1203 13:04:49.293675 4990 generic.go:334] "Generic (PLEG): container finished" podID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerID="3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2" exitCode=0 Dec 03 13:04:49 crc kubenswrapper[4990]: I1203 13:04:49.293771 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerDied","Data":"3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2"} Dec 03 13:04:49 crc kubenswrapper[4990]: I1203 13:04:49.296159 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" event={"ID":"32c217f2-3a1b-47f5-91d9-b6634935c9f0","Type":"ContainerStarted","Data":"e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230"} Dec 03 13:04:49 crc kubenswrapper[4990]: I1203 13:04:49.296205 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" event={"ID":"32c217f2-3a1b-47f5-91d9-b6634935c9f0","Type":"ContainerStarted","Data":"7f6ba03adb981fe163b06d93244df2469be04420e13132acee6fee0bc8012157"} Dec 03 13:04:49 crc kubenswrapper[4990]: I1203 13:04:49.568360 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:50 crc kubenswrapper[4990]: I1203 13:04:50.052150 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:50 crc kubenswrapper[4990]: I1203 13:04:50.306864 4990 generic.go:334] "Generic (PLEG): container finished" podID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerID="e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230" exitCode=0 Dec 03 13:04:50 crc kubenswrapper[4990]: I1203 13:04:50.306971 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" event={"ID":"32c217f2-3a1b-47f5-91d9-b6634935c9f0","Type":"ContainerDied","Data":"e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230"} Dec 03 13:04:50 crc kubenswrapper[4990]: I1203 13:04:50.308077 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-log" containerID="cri-o://c02236eb6c3d6123a966261adcb0298733d380099030b20571666800bee2441e" gracePeriod=30 Dec 03 13:04:50 crc kubenswrapper[4990]: I1203 13:04:50.308134 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-api" containerID="cri-o://d640fdf5618d25d7277da5d497e23562a7c88e997a561356044c85aa0c949b78" gracePeriod=30 Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.319981 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerStarted","Data":"2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434"} Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.322352 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" event={"ID":"32c217f2-3a1b-47f5-91d9-b6634935c9f0","Type":"ContainerStarted","Data":"1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023"} Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.323159 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.324969 4990 generic.go:334] "Generic (PLEG): container finished" podID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerID="c02236eb6c3d6123a966261adcb0298733d380099030b20571666800bee2441e" exitCode=143 Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.324997 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce28826a-46d4-4ccb-8f2f-65cd0f035c36","Type":"ContainerDied","Data":"c02236eb6c3d6123a966261adcb0298733d380099030b20571666800bee2441e"} Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.345016 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pg5vd" podStartSLOduration=3.066231846 podStartE2EDuration="6.344997855s" podCreationTimestamp="2025-12-03 13:04:45 +0000 UTC" firstStartedPulling="2025-12-03 13:04:47.252642893 +0000 UTC m=+1635.394554122" lastFinishedPulling="2025-12-03 13:04:50.531408902 +0000 UTC m=+1638.673320131" observedRunningTime="2025-12-03 13:04:51.340074174 +0000 UTC m=+1639.481985403" watchObservedRunningTime="2025-12-03 13:04:51.344997855 +0000 UTC m=+1639.486909074" Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.367044 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" podStartSLOduration=4.367027718 podStartE2EDuration="4.367027718s" podCreationTimestamp="2025-12-03 13:04:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:51.358161764 +0000 UTC m=+1639.500073003" watchObservedRunningTime="2025-12-03 13:04:51.367027718 +0000 UTC m=+1639.508938937" Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.464269 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 13:04:51 crc kubenswrapper[4990]: I1203 13:04:51.985155 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:52 crc kubenswrapper[4990]: I1203 13:04:52.331521 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-central-agent" containerID="cri-o://697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea" gracePeriod=30 Dec 03 13:04:52 crc kubenswrapper[4990]: I1203 13:04:52.333140 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="proxy-httpd" containerID="cri-o://f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261" gracePeriod=30 Dec 03 13:04:52 crc kubenswrapper[4990]: I1203 13:04:52.333289 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="sg-core" containerID="cri-o://ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006" gracePeriod=30 Dec 03 13:04:52 crc kubenswrapper[4990]: I1203 13:04:52.333403 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-notification-agent" containerID="cri-o://ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305" gracePeriod=30 Dec 03 13:04:53 crc kubenswrapper[4990]: I1203 13:04:53.376968 4990 generic.go:334] "Generic (PLEG): container finished" podID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerID="f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261" exitCode=0 Dec 03 13:04:53 crc kubenswrapper[4990]: I1203 13:04:53.377054 4990 generic.go:334] "Generic (PLEG): container finished" podID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerID="ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006" exitCode=2 Dec 03 13:04:53 crc kubenswrapper[4990]: I1203 13:04:53.377066 4990 generic.go:334] "Generic (PLEG): container finished" podID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerID="697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea" exitCode=0 Dec 03 13:04:53 crc kubenswrapper[4990]: I1203 13:04:53.378053 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerDied","Data":"f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261"} Dec 03 13:04:53 crc kubenswrapper[4990]: I1203 13:04:53.378084 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerDied","Data":"ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006"} Dec 03 13:04:53 crc kubenswrapper[4990]: I1203 13:04:53.378099 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerDied","Data":"697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea"} Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.388157 4990 generic.go:334] "Generic (PLEG): container finished" podID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerID="d640fdf5618d25d7277da5d497e23562a7c88e997a561356044c85aa0c949b78" exitCode=0 Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.388234 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce28826a-46d4-4ccb-8f2f-65cd0f035c36","Type":"ContainerDied","Data":"d640fdf5618d25d7277da5d497e23562a7c88e997a561356044c85aa0c949b78"} Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.388628 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce28826a-46d4-4ccb-8f2f-65cd0f035c36","Type":"ContainerDied","Data":"bfc16465b787828760188d59779febd349fc0f5561f24907d791086a247345b2"} Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.388651 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfc16465b787828760188d59779febd349fc0f5561f24907d791086a247345b2" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.426437 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.568393 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.589074 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.591720 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-config-data\") pod \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.591915 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-logs\") pod \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.591964 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-combined-ca-bundle\") pod \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.591997 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsdfc\" (UniqueName: \"kubernetes.io/projected/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-kube-api-access-lsdfc\") pod \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\" (UID: \"ce28826a-46d4-4ccb-8f2f-65cd0f035c36\") " Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.592577 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-logs" (OuterVolumeSpecName: "logs") pod "ce28826a-46d4-4ccb-8f2f-65cd0f035c36" (UID: "ce28826a-46d4-4ccb-8f2f-65cd0f035c36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.600146 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-kube-api-access-lsdfc" (OuterVolumeSpecName: "kube-api-access-lsdfc") pod "ce28826a-46d4-4ccb-8f2f-65cd0f035c36" (UID: "ce28826a-46d4-4ccb-8f2f-65cd0f035c36"). InnerVolumeSpecName "kube-api-access-lsdfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.633717 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-config-data" (OuterVolumeSpecName: "config-data") pod "ce28826a-46d4-4ccb-8f2f-65cd0f035c36" (UID: "ce28826a-46d4-4ccb-8f2f-65cd0f035c36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.647666 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce28826a-46d4-4ccb-8f2f-65cd0f035c36" (UID: "ce28826a-46d4-4ccb-8f2f-65cd0f035c36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.694481 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.694699 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.694787 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:54 crc kubenswrapper[4990]: I1203 13:04:54.694857 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsdfc\" (UniqueName: \"kubernetes.io/projected/ce28826a-46d4-4ccb-8f2f-65cd0f035c36-kube-api-access-lsdfc\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.396137 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.415576 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.453919 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.462422 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.480223 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:55 crc kubenswrapper[4990]: E1203 13:04:55.481808 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-api" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.481937 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-api" Dec 03 13:04:55 crc kubenswrapper[4990]: E1203 13:04:55.482012 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-log" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.482157 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-log" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.482392 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-api" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.482601 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" containerName="nova-api-log" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.483695 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.494558 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.494789 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.494915 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.507142 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.519301 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.519539 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.531807 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d459c9c-7ea1-42c1-996f-85f099f16e90-logs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.531868 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-config-data\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.531900 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.532003 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.532731 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgbtl\" (UniqueName: \"kubernetes.io/projected/0d459c9c-7ea1-42c1-996f-85f099f16e90-kube-api-access-bgbtl\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.532987 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-public-tls-certs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.595950 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.634654 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d459c9c-7ea1-42c1-996f-85f099f16e90-logs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.635009 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-config-data\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.635119 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.635254 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.635404 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgbtl\" (UniqueName: \"kubernetes.io/projected/0d459c9c-7ea1-42c1-996f-85f099f16e90-kube-api-access-bgbtl\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.635601 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-public-tls-certs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.638292 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d459c9c-7ea1-42c1-996f-85f099f16e90-logs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.640640 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-config-data\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.645027 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-public-tls-certs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.677907 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-zgltr"] Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.681385 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.695481 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.695898 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.696250 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-zgltr"] Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.702378 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.703524 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgbtl\" (UniqueName: \"kubernetes.io/projected/0d459c9c-7ea1-42c1-996f-85f099f16e90-kube-api-access-bgbtl\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.710127 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.824916 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.853058 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.853181 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65gvp\" (UniqueName: \"kubernetes.io/projected/ae0b69fc-aee2-444e-888d-c18a35a386b5-kube-api-access-65gvp\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.853217 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-config-data\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.853345 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-scripts\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.971234 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.971339 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65gvp\" (UniqueName: \"kubernetes.io/projected/ae0b69fc-aee2-444e-888d-c18a35a386b5-kube-api-access-65gvp\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.971368 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-config-data\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.971504 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-scripts\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.987053 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-config-data\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.987075 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.990734 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-scripts\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:55 crc kubenswrapper[4990]: I1203 13:04:55.997143 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65gvp\" (UniqueName: \"kubernetes.io/projected/ae0b69fc-aee2-444e-888d-c18a35a386b5-kube-api-access-65gvp\") pod \"nova-cell1-cell-mapping-zgltr\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.097953 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.139714 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278116 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcbbw\" (UniqueName: \"kubernetes.io/projected/87c13d56-b975-4be0-94f5-efa2e80b4b31-kube-api-access-gcbbw\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278381 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-run-httpd\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278401 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-scripts\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278467 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-config-data\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278521 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-ceilometer-tls-certs\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278617 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-combined-ca-bundle\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278724 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-sg-core-conf-yaml\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.278767 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-log-httpd\") pod \"87c13d56-b975-4be0-94f5-efa2e80b4b31\" (UID: \"87c13d56-b975-4be0-94f5-efa2e80b4b31\") " Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.279565 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.280327 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.283980 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce28826a-46d4-4ccb-8f2f-65cd0f035c36" path="/var/lib/kubelet/pods/ce28826a-46d4-4ccb-8f2f-65cd0f035c36/volumes" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.284133 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-scripts" (OuterVolumeSpecName: "scripts") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.294238 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c13d56-b975-4be0-94f5-efa2e80b4b31-kube-api-access-gcbbw" (OuterVolumeSpecName: "kube-api-access-gcbbw") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "kube-api-access-gcbbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.317683 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.362106 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.362157 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.373300 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380690 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380716 4990 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380727 4990 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380735 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcbbw\" (UniqueName: \"kubernetes.io/projected/87c13d56-b975-4be0-94f5-efa2e80b4b31-kube-api-access-gcbbw\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380744 4990 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c13d56-b975-4be0-94f5-efa2e80b4b31-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380752 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.380760 4990 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.412344 4990 generic.go:334] "Generic (PLEG): container finished" podID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerID="ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305" exitCode=0 Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.413665 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerDied","Data":"ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305"} Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.413716 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c13d56-b975-4be0-94f5-efa2e80b4b31","Type":"ContainerDied","Data":"cf4dc51eff43cdb7eec0f0dc94c7c7da9011fb7dedca2e1ed57438dbcca477c2"} Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.413736 4990 scope.go:117] "RemoveContainer" containerID="f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.413685 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.418579 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d459c9c-7ea1-42c1-996f-85f099f16e90","Type":"ContainerStarted","Data":"29280ed8487092f866a3a14a250790dd3e312a9eab4ff70b8fcd0da6a4032728"} Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.429081 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-config-data" (OuterVolumeSpecName: "config-data") pod "87c13d56-b975-4be0-94f5-efa2e80b4b31" (UID: "87c13d56-b975-4be0-94f5-efa2e80b4b31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.470292 4990 scope.go:117] "RemoveContainer" containerID="ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.482971 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c13d56-b975-4be0-94f5-efa2e80b4b31-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.494680 4990 scope.go:117] "RemoveContainer" containerID="ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.494859 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.528961 4990 scope.go:117] "RemoveContainer" containerID="697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.552649 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pg5vd"] Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.555106 4990 scope.go:117] "RemoveContainer" containerID="f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.555562 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261\": container with ID starting with f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261 not found: ID does not exist" containerID="f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.555609 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261"} err="failed to get container status \"f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261\": rpc error: code = NotFound desc = could not find container \"f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261\": container with ID starting with f65215e1e10edc2a32c5a0ab670cfb97836009f88ede6a0d347d646668e69261 not found: ID does not exist" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.555641 4990 scope.go:117] "RemoveContainer" containerID="ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.556359 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006\": container with ID starting with ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006 not found: ID does not exist" containerID="ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.556474 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006"} err="failed to get container status \"ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006\": rpc error: code = NotFound desc = could not find container \"ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006\": container with ID starting with ec74169b8bd883229f02a32edf87012e6bfc640051af9cfcf676ba424b6e1006 not found: ID does not exist" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.556559 4990 scope.go:117] "RemoveContainer" containerID="ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.556997 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305\": container with ID starting with ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305 not found: ID does not exist" containerID="ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.557094 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305"} err="failed to get container status \"ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305\": rpc error: code = NotFound desc = could not find container \"ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305\": container with ID starting with ab7f6ca3db32cc6fe6cc3f674190a751c71b64cc5ffcf9c5bc60b15d35da1305 not found: ID does not exist" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.557160 4990 scope.go:117] "RemoveContainer" containerID="697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.557977 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea\": container with ID starting with 697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea not found: ID does not exist" containerID="697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.558063 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea"} err="failed to get container status \"697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea\": rpc error: code = NotFound desc = could not find container \"697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea\": container with ID starting with 697b07a22c4590ed5d2ff69bf033fc4b4142a1b87c209966568f93afa233c3ea not found: ID does not exist" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.593645 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-zgltr"] Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.765375 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.780509 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.798556 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.800484 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-notification-agent" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800512 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-notification-agent" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.800546 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="proxy-httpd" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800554 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="proxy-httpd" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.800591 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="sg-core" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800598 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="sg-core" Dec 03 13:04:56 crc kubenswrapper[4990]: E1203 13:04:56.800613 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-central-agent" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800620 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-central-agent" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800831 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="proxy-httpd" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800860 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-notification-agent" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800871 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="sg-core" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.800884 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" containerName="ceilometer-central-agent" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.803113 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.810792 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.810821 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.811771 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.814639 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891576 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891685 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cg9k\" (UniqueName: \"kubernetes.io/projected/901bdc2a-da65-4388-9865-67a1ff3eec17-kube-api-access-4cg9k\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891811 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-config-data\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891833 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-scripts\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891858 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891885 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891910 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/901bdc2a-da65-4388-9865-67a1ff3eec17-log-httpd\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.891968 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/901bdc2a-da65-4388-9865-67a1ff3eec17-run-httpd\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.994012 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cg9k\" (UniqueName: \"kubernetes.io/projected/901bdc2a-da65-4388-9865-67a1ff3eec17-kube-api-access-4cg9k\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.994405 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-config-data\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.994529 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-scripts\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.994633 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.994746 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.994861 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/901bdc2a-da65-4388-9865-67a1ff3eec17-log-httpd\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.995004 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/901bdc2a-da65-4388-9865-67a1ff3eec17-run-httpd\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.995123 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.995401 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/901bdc2a-da65-4388-9865-67a1ff3eec17-log-httpd\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:56 crc kubenswrapper[4990]: I1203 13:04:56.995424 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/901bdc2a-da65-4388-9865-67a1ff3eec17-run-httpd\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.000084 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-scripts\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.000680 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.001296 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.001383 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-config-data\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.003258 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/901bdc2a-da65-4388-9865-67a1ff3eec17-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.023312 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cg9k\" (UniqueName: \"kubernetes.io/projected/901bdc2a-da65-4388-9865-67a1ff3eec17-kube-api-access-4cg9k\") pod \"ceilometer-0\" (UID: \"901bdc2a-da65-4388-9865-67a1ff3eec17\") " pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.125792 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.435973 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d459c9c-7ea1-42c1-996f-85f099f16e90","Type":"ContainerStarted","Data":"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502"} Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.436276 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d459c9c-7ea1-42c1-996f-85f099f16e90","Type":"ContainerStarted","Data":"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14"} Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.439390 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zgltr" event={"ID":"ae0b69fc-aee2-444e-888d-c18a35a386b5","Type":"ContainerStarted","Data":"6904368d80805855204b8af907c9324de9ad1994c0475ee23365de506dcab7c1"} Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.439429 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zgltr" event={"ID":"ae0b69fc-aee2-444e-888d-c18a35a386b5","Type":"ContainerStarted","Data":"e0ef19a982a9e68f9957e64a412f6fdf4327c908dc9a620bdbe13cfef6b7a569"} Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.459920 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.459901992 podStartE2EDuration="2.459901992s" podCreationTimestamp="2025-12-03 13:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:57.455721681 +0000 UTC m=+1645.597632910" watchObservedRunningTime="2025-12-03 13:04:57.459901992 +0000 UTC m=+1645.601813221" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.482417 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-zgltr" podStartSLOduration=2.482396878 podStartE2EDuration="2.482396878s" podCreationTimestamp="2025-12-03 13:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:04:57.474089158 +0000 UTC m=+1645.616000387" watchObservedRunningTime="2025-12-03 13:04:57.482396878 +0000 UTC m=+1645.624308097" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.627267 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.808727 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.881803 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-jngsh"] Dec 03 13:04:57 crc kubenswrapper[4990]: I1203 13:04:57.883341 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerName="dnsmasq-dns" containerID="cri-o://e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423" gracePeriod=10 Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.291416 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c13d56-b975-4be0-94f5-efa2e80b4b31" path="/var/lib/kubelet/pods/87c13d56-b975-4be0-94f5-efa2e80b4b31/volumes" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.369816 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.449124 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"901bdc2a-da65-4388-9865-67a1ff3eec17","Type":"ContainerStarted","Data":"91e12b5e26faf8f680909d630eb7f68f436e443b9188126ae1496861f2334222"} Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.457135 4990 generic.go:334] "Generic (PLEG): container finished" podID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerID="e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423" exitCode=0 Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.457431 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pg5vd" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="registry-server" containerID="cri-o://2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434" gracePeriod=2 Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.457882 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.457991 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" event={"ID":"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5","Type":"ContainerDied","Data":"e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423"} Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.458052 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-jngsh" event={"ID":"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5","Type":"ContainerDied","Data":"77e6c024bcf9a88bd785e3abef9d9b7285b393b1569289b7a07271def542f508"} Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.458072 4990 scope.go:117] "RemoveContainer" containerID="e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.504843 4990 scope.go:117] "RemoveContainer" containerID="13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.530093 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfp9g\" (UniqueName: \"kubernetes.io/projected/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-kube-api-access-xfp9g\") pod \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.530211 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-swift-storage-0\") pod \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.530310 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-nb\") pod \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.530397 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-svc\") pod \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.530422 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-config\") pod \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.530441 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-sb\") pod \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\" (UID: \"107ae5ef-f132-4fc5-a9d7-b9b649dd98d5\") " Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.535979 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-kube-api-access-xfp9g" (OuterVolumeSpecName: "kube-api-access-xfp9g") pod "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" (UID: "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5"). InnerVolumeSpecName "kube-api-access-xfp9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.571973 4990 scope.go:117] "RemoveContainer" containerID="e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423" Dec 03 13:04:58 crc kubenswrapper[4990]: E1203 13:04:58.572435 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423\": container with ID starting with e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423 not found: ID does not exist" containerID="e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.572488 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423"} err="failed to get container status \"e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423\": rpc error: code = NotFound desc = could not find container \"e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423\": container with ID starting with e63ed5be8d020772630e6138e94a6a184ed22d0680aee681c55cc559ac5d9423 not found: ID does not exist" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.572515 4990 scope.go:117] "RemoveContainer" containerID="13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d" Dec 03 13:04:58 crc kubenswrapper[4990]: E1203 13:04:58.572866 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d\": container with ID starting with 13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d not found: ID does not exist" containerID="13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.572898 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d"} err="failed to get container status \"13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d\": rpc error: code = NotFound desc = could not find container \"13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d\": container with ID starting with 13e44b9a9880e9d376436b8782c7266cd0679a9b156c8053d97d385ebf60522d not found: ID does not exist" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.582961 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" (UID: "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.586830 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-config" (OuterVolumeSpecName: "config") pod "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" (UID: "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.586965 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" (UID: "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.599503 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" (UID: "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.602996 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" (UID: "107ae5ef-f132-4fc5-a9d7-b9b649dd98d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.633265 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.633306 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.633319 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.633329 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.633341 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.633353 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfp9g\" (UniqueName: \"kubernetes.io/projected/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5-kube-api-access-xfp9g\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.811569 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-jngsh"] Dec 03 13:04:58 crc kubenswrapper[4990]: I1203 13:04:58.822998 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-jngsh"] Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.178568 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.248082 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-utilities\") pod \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.248325 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26rjq\" (UniqueName: \"kubernetes.io/projected/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-kube-api-access-26rjq\") pod \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.248443 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-catalog-content\") pod \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\" (UID: \"c684cd85-ef9f-448b-bc2d-daaf12bf27f4\") " Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.249066 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-utilities" (OuterVolumeSpecName: "utilities") pod "c684cd85-ef9f-448b-bc2d-daaf12bf27f4" (UID: "c684cd85-ef9f-448b-bc2d-daaf12bf27f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.255597 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-kube-api-access-26rjq" (OuterVolumeSpecName: "kube-api-access-26rjq") pod "c684cd85-ef9f-448b-bc2d-daaf12bf27f4" (UID: "c684cd85-ef9f-448b-bc2d-daaf12bf27f4"). InnerVolumeSpecName "kube-api-access-26rjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.350944 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26rjq\" (UniqueName: \"kubernetes.io/projected/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-kube-api-access-26rjq\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.350987 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.374423 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c684cd85-ef9f-448b-bc2d-daaf12bf27f4" (UID: "c684cd85-ef9f-448b-bc2d-daaf12bf27f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.452858 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c684cd85-ef9f-448b-bc2d-daaf12bf27f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.468756 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"901bdc2a-da65-4388-9865-67a1ff3eec17","Type":"ContainerStarted","Data":"bf4fce966f2b917f3772003f5a4dbcdefa0d74556db8f5f03627919f137552da"} Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.468822 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"901bdc2a-da65-4388-9865-67a1ff3eec17","Type":"ContainerStarted","Data":"2a57c04edf46d576de8d3ec8c1c73c1c396e37cc672c77419a628f2c3fe8e55e"} Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.473264 4990 generic.go:334] "Generic (PLEG): container finished" podID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerID="2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434" exitCode=0 Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.473321 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pg5vd" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.473345 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerDied","Data":"2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434"} Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.473619 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pg5vd" event={"ID":"c684cd85-ef9f-448b-bc2d-daaf12bf27f4","Type":"ContainerDied","Data":"90a05cbc2a9812dd3f7c8d0694131f4b0e8112374f6d4b2652dbe4cbd4e8d312"} Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.473642 4990 scope.go:117] "RemoveContainer" containerID="2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.507560 4990 scope.go:117] "RemoveContainer" containerID="3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.513930 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pg5vd"] Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.526020 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pg5vd"] Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.539093 4990 scope.go:117] "RemoveContainer" containerID="3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.568177 4990 scope.go:117] "RemoveContainer" containerID="2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434" Dec 03 13:04:59 crc kubenswrapper[4990]: E1203 13:04:59.568932 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434\": container with ID starting with 2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434 not found: ID does not exist" containerID="2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.568988 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434"} err="failed to get container status \"2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434\": rpc error: code = NotFound desc = could not find container \"2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434\": container with ID starting with 2805a0d006ccaca4504d9fc80246179920575e9e498fc98d37e66f3827394434 not found: ID does not exist" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.569020 4990 scope.go:117] "RemoveContainer" containerID="3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2" Dec 03 13:04:59 crc kubenswrapper[4990]: E1203 13:04:59.569621 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2\": container with ID starting with 3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2 not found: ID does not exist" containerID="3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.569657 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2"} err="failed to get container status \"3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2\": rpc error: code = NotFound desc = could not find container \"3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2\": container with ID starting with 3541da21480f9c9819f1cc33e3b69c0e0e9a4426e05734eea824f3fbbd4abda2 not found: ID does not exist" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.569683 4990 scope.go:117] "RemoveContainer" containerID="3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11" Dec 03 13:04:59 crc kubenswrapper[4990]: E1203 13:04:59.569975 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11\": container with ID starting with 3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11 not found: ID does not exist" containerID="3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11" Dec 03 13:04:59 crc kubenswrapper[4990]: I1203 13:04:59.570010 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11"} err="failed to get container status \"3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11\": rpc error: code = NotFound desc = could not find container \"3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11\": container with ID starting with 3713892aca65ed4a0268f412d155525dbe3433514d2fc41cf48b855cbb23de11 not found: ID does not exist" Dec 03 13:05:00 crc kubenswrapper[4990]: I1203 13:05:00.277399 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" path="/var/lib/kubelet/pods/107ae5ef-f132-4fc5-a9d7-b9b649dd98d5/volumes" Dec 03 13:05:00 crc kubenswrapper[4990]: I1203 13:05:00.279116 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" path="/var/lib/kubelet/pods/c684cd85-ef9f-448b-bc2d-daaf12bf27f4/volumes" Dec 03 13:05:00 crc kubenswrapper[4990]: I1203 13:05:00.490889 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"901bdc2a-da65-4388-9865-67a1ff3eec17","Type":"ContainerStarted","Data":"656f7e42e05b98aa22e2b7aee8571048ffe7414f1a41c8960accdc70b1b53bd1"} Dec 03 13:05:03 crc kubenswrapper[4990]: I1203 13:05:03.265393 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:05:03 crc kubenswrapper[4990]: E1203 13:05:03.266772 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:05:03 crc kubenswrapper[4990]: I1203 13:05:03.522924 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"901bdc2a-da65-4388-9865-67a1ff3eec17","Type":"ContainerStarted","Data":"2078595a8063b91d9371c3a1a11353249b8d8edd8a9c420613087d7a51c222ba"} Dec 03 13:05:03 crc kubenswrapper[4990]: I1203 13:05:03.523393 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 13:05:03 crc kubenswrapper[4990]: I1203 13:05:03.524553 4990 generic.go:334] "Generic (PLEG): container finished" podID="ae0b69fc-aee2-444e-888d-c18a35a386b5" containerID="6904368d80805855204b8af907c9324de9ad1994c0475ee23365de506dcab7c1" exitCode=0 Dec 03 13:05:03 crc kubenswrapper[4990]: I1203 13:05:03.524708 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zgltr" event={"ID":"ae0b69fc-aee2-444e-888d-c18a35a386b5","Type":"ContainerDied","Data":"6904368d80805855204b8af907c9324de9ad1994c0475ee23365de506dcab7c1"} Dec 03 13:05:03 crc kubenswrapper[4990]: I1203 13:05:03.545311 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.704227591 podStartE2EDuration="7.545290096s" podCreationTimestamp="2025-12-03 13:04:56 +0000 UTC" firstStartedPulling="2025-12-03 13:04:57.628715066 +0000 UTC m=+1645.770626295" lastFinishedPulling="2025-12-03 13:05:02.469777561 +0000 UTC m=+1650.611688800" observedRunningTime="2025-12-03 13:05:03.541889836 +0000 UTC m=+1651.683801065" watchObservedRunningTime="2025-12-03 13:05:03.545290096 +0000 UTC m=+1651.687201325" Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.902188 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.975630 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-config-data\") pod \"ae0b69fc-aee2-444e-888d-c18a35a386b5\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.976309 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-scripts\") pod \"ae0b69fc-aee2-444e-888d-c18a35a386b5\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.976351 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65gvp\" (UniqueName: \"kubernetes.io/projected/ae0b69fc-aee2-444e-888d-c18a35a386b5-kube-api-access-65gvp\") pod \"ae0b69fc-aee2-444e-888d-c18a35a386b5\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.976393 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-combined-ca-bundle\") pod \"ae0b69fc-aee2-444e-888d-c18a35a386b5\" (UID: \"ae0b69fc-aee2-444e-888d-c18a35a386b5\") " Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.981662 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-scripts" (OuterVolumeSpecName: "scripts") pod "ae0b69fc-aee2-444e-888d-c18a35a386b5" (UID: "ae0b69fc-aee2-444e-888d-c18a35a386b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:04 crc kubenswrapper[4990]: I1203 13:05:04.982467 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae0b69fc-aee2-444e-888d-c18a35a386b5-kube-api-access-65gvp" (OuterVolumeSpecName: "kube-api-access-65gvp") pod "ae0b69fc-aee2-444e-888d-c18a35a386b5" (UID: "ae0b69fc-aee2-444e-888d-c18a35a386b5"). InnerVolumeSpecName "kube-api-access-65gvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.004870 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-config-data" (OuterVolumeSpecName: "config-data") pod "ae0b69fc-aee2-444e-888d-c18a35a386b5" (UID: "ae0b69fc-aee2-444e-888d-c18a35a386b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.013684 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae0b69fc-aee2-444e-888d-c18a35a386b5" (UID: "ae0b69fc-aee2-444e-888d-c18a35a386b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.078818 4990 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.078867 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65gvp\" (UniqueName: \"kubernetes.io/projected/ae0b69fc-aee2-444e-888d-c18a35a386b5-kube-api-access-65gvp\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.078896 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.078926 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0b69fc-aee2-444e-888d-c18a35a386b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.545329 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zgltr" event={"ID":"ae0b69fc-aee2-444e-888d-c18a35a386b5","Type":"ContainerDied","Data":"e0ef19a982a9e68f9957e64a412f6fdf4327c908dc9a620bdbe13cfef6b7a569"} Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.545368 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0ef19a982a9e68f9957e64a412f6fdf4327c908dc9a620bdbe13cfef6b7a569" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.545387 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zgltr" Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.743006 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.743328 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-log" containerID="cri-o://6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14" gracePeriod=30 Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.743378 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-api" containerID="cri-o://f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502" gracePeriod=30 Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.761701 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.763406 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="32409013-6281-4048-8606-ae246d1b1aed" containerName="nova-scheduler-scheduler" containerID="cri-o://ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf" gracePeriod=30 Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.788530 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.788791 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-log" containerID="cri-o://df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146" gracePeriod=30 Dec 03 13:05:05 crc kubenswrapper[4990]: I1203 13:05:05.789266 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-metadata" containerID="cri-o://bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e" gracePeriod=30 Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.335709 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.500903 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-public-tls-certs\") pod \"0d459c9c-7ea1-42c1-996f-85f099f16e90\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.500976 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-combined-ca-bundle\") pod \"0d459c9c-7ea1-42c1-996f-85f099f16e90\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.501039 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-internal-tls-certs\") pod \"0d459c9c-7ea1-42c1-996f-85f099f16e90\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.501088 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d459c9c-7ea1-42c1-996f-85f099f16e90-logs\") pod \"0d459c9c-7ea1-42c1-996f-85f099f16e90\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.501115 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-config-data\") pod \"0d459c9c-7ea1-42c1-996f-85f099f16e90\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.501533 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d459c9c-7ea1-42c1-996f-85f099f16e90-logs" (OuterVolumeSpecName: "logs") pod "0d459c9c-7ea1-42c1-996f-85f099f16e90" (UID: "0d459c9c-7ea1-42c1-996f-85f099f16e90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.501566 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgbtl\" (UniqueName: \"kubernetes.io/projected/0d459c9c-7ea1-42c1-996f-85f099f16e90-kube-api-access-bgbtl\") pod \"0d459c9c-7ea1-42c1-996f-85f099f16e90\" (UID: \"0d459c9c-7ea1-42c1-996f-85f099f16e90\") " Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.502019 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d459c9c-7ea1-42c1-996f-85f099f16e90-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.510489 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d459c9c-7ea1-42c1-996f-85f099f16e90-kube-api-access-bgbtl" (OuterVolumeSpecName: "kube-api-access-bgbtl") pod "0d459c9c-7ea1-42c1-996f-85f099f16e90" (UID: "0d459c9c-7ea1-42c1-996f-85f099f16e90"). InnerVolumeSpecName "kube-api-access-bgbtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.528137 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-config-data" (OuterVolumeSpecName: "config-data") pod "0d459c9c-7ea1-42c1-996f-85f099f16e90" (UID: "0d459c9c-7ea1-42c1-996f-85f099f16e90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.539586 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d459c9c-7ea1-42c1-996f-85f099f16e90" (UID: "0d459c9c-7ea1-42c1-996f-85f099f16e90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561681 4990 generic.go:334] "Generic (PLEG): container finished" podID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerID="f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502" exitCode=0 Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561712 4990 generic.go:334] "Generic (PLEG): container finished" podID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerID="6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14" exitCode=143 Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561759 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d459c9c-7ea1-42c1-996f-85f099f16e90","Type":"ContainerDied","Data":"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502"} Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561789 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d459c9c-7ea1-42c1-996f-85f099f16e90","Type":"ContainerDied","Data":"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14"} Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561800 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d459c9c-7ea1-42c1-996f-85f099f16e90","Type":"ContainerDied","Data":"29280ed8487092f866a3a14a250790dd3e312a9eab4ff70b8fcd0da6a4032728"} Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561814 4990 scope.go:117] "RemoveContainer" containerID="f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561943 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.561800 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0d459c9c-7ea1-42c1-996f-85f099f16e90" (UID: "0d459c9c-7ea1-42c1-996f-85f099f16e90"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.562144 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0d459c9c-7ea1-42c1-996f-85f099f16e90" (UID: "0d459c9c-7ea1-42c1-996f-85f099f16e90"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.567214 4990 generic.go:334] "Generic (PLEG): container finished" podID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerID="df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146" exitCode=143 Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.567254 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0780aac2-023d-4f4a-8fef-ccbaf8af60c7","Type":"ContainerDied","Data":"df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146"} Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.604232 4990 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.604278 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.604290 4990 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.604305 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d459c9c-7ea1-42c1-996f-85f099f16e90-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.604317 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgbtl\" (UniqueName: \"kubernetes.io/projected/0d459c9c-7ea1-42c1-996f-85f099f16e90-kube-api-access-bgbtl\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.609466 4990 scope.go:117] "RemoveContainer" containerID="6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.636361 4990 scope.go:117] "RemoveContainer" containerID="f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502" Dec 03 13:05:06 crc kubenswrapper[4990]: E1203 13:05:06.636872 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502\": container with ID starting with f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502 not found: ID does not exist" containerID="f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.636915 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502"} err="failed to get container status \"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502\": rpc error: code = NotFound desc = could not find container \"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502\": container with ID starting with f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502 not found: ID does not exist" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.636943 4990 scope.go:117] "RemoveContainer" containerID="6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14" Dec 03 13:05:06 crc kubenswrapper[4990]: E1203 13:05:06.637344 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14\": container with ID starting with 6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14 not found: ID does not exist" containerID="6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.637382 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14"} err="failed to get container status \"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14\": rpc error: code = NotFound desc = could not find container \"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14\": container with ID starting with 6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14 not found: ID does not exist" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.637415 4990 scope.go:117] "RemoveContainer" containerID="f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.637788 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502"} err="failed to get container status \"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502\": rpc error: code = NotFound desc = could not find container \"f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502\": container with ID starting with f5dc0b91d174dac6977024eb12676d1d5feaff851331d11d338f70d2c9cce502 not found: ID does not exist" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.637821 4990 scope.go:117] "RemoveContainer" containerID="6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.638167 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14"} err="failed to get container status \"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14\": rpc error: code = NotFound desc = could not find container \"6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14\": container with ID starting with 6878d7ff29e2d76c4c8bacd87d90c395580a561a0cb6f711ca3440ababb99e14 not found: ID does not exist" Dec 03 13:05:06 crc kubenswrapper[4990]: I1203 13:05:06.999781 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.021041 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.032489 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071029 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071499 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerName="dnsmasq-dns" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071518 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerName="dnsmasq-dns" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071535 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32409013-6281-4048-8606-ae246d1b1aed" containerName="nova-scheduler-scheduler" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071542 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="32409013-6281-4048-8606-ae246d1b1aed" containerName="nova-scheduler-scheduler" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071550 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-api" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071556 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-api" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071571 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0b69fc-aee2-444e-888d-c18a35a386b5" containerName="nova-manage" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071577 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0b69fc-aee2-444e-888d-c18a35a386b5" containerName="nova-manage" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071588 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-log" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071594 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-log" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071614 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="registry-server" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071620 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="registry-server" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071631 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerName="init" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071637 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerName="init" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071646 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="extract-utilities" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071652 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="extract-utilities" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.071674 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="extract-content" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071679 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="extract-content" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071841 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="32409013-6281-4048-8606-ae246d1b1aed" containerName="nova-scheduler-scheduler" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071852 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="107ae5ef-f132-4fc5-a9d7-b9b649dd98d5" containerName="dnsmasq-dns" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071867 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-api" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071880 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" containerName="nova-api-log" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071898 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c684cd85-ef9f-448b-bc2d-daaf12bf27f4" containerName="registry-server" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.071911 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0b69fc-aee2-444e-888d-c18a35a386b5" containerName="nova-manage" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.072897 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.075185 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.075228 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.075613 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.098642 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.113354 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-combined-ca-bundle\") pod \"32409013-6281-4048-8606-ae246d1b1aed\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.113466 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrqph\" (UniqueName: \"kubernetes.io/projected/32409013-6281-4048-8606-ae246d1b1aed-kube-api-access-wrqph\") pod \"32409013-6281-4048-8606-ae246d1b1aed\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.113521 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-config-data\") pod \"32409013-6281-4048-8606-ae246d1b1aed\" (UID: \"32409013-6281-4048-8606-ae246d1b1aed\") " Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.121532 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32409013-6281-4048-8606-ae246d1b1aed-kube-api-access-wrqph" (OuterVolumeSpecName: "kube-api-access-wrqph") pod "32409013-6281-4048-8606-ae246d1b1aed" (UID: "32409013-6281-4048-8606-ae246d1b1aed"). InnerVolumeSpecName "kube-api-access-wrqph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.148023 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32409013-6281-4048-8606-ae246d1b1aed" (UID: "32409013-6281-4048-8606-ae246d1b1aed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.172123 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-config-data" (OuterVolumeSpecName: "config-data") pod "32409013-6281-4048-8606-ae246d1b1aed" (UID: "32409013-6281-4048-8606-ae246d1b1aed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.215728 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-config-data\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.215816 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.215882 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.216141 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.216219 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd7jt\" (UniqueName: \"kubernetes.io/projected/c704a76d-fc53-4741-9bd5-b7893e1c96a5-kube-api-access-vd7jt\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.216365 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c704a76d-fc53-4741-9bd5-b7893e1c96a5-logs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.216647 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.216673 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrqph\" (UniqueName: \"kubernetes.io/projected/32409013-6281-4048-8606-ae246d1b1aed-kube-api-access-wrqph\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.216689 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32409013-6281-4048-8606-ae246d1b1aed-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.317887 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-config-data\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.317945 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.317993 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.318035 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.318052 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd7jt\" (UniqueName: \"kubernetes.io/projected/c704a76d-fc53-4741-9bd5-b7893e1c96a5-kube-api-access-vd7jt\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.318091 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c704a76d-fc53-4741-9bd5-b7893e1c96a5-logs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.318474 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c704a76d-fc53-4741-9bd5-b7893e1c96a5-logs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.321549 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-config-data\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.322070 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.323755 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.325280 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c704a76d-fc53-4741-9bd5-b7893e1c96a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.341726 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd7jt\" (UniqueName: \"kubernetes.io/projected/c704a76d-fc53-4741-9bd5-b7893e1c96a5-kube-api-access-vd7jt\") pod \"nova-api-0\" (UID: \"c704a76d-fc53-4741-9bd5-b7893e1c96a5\") " pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.405647 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.585321 4990 generic.go:334] "Generic (PLEG): container finished" podID="32409013-6281-4048-8606-ae246d1b1aed" containerID="ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf" exitCode=0 Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.585370 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32409013-6281-4048-8606-ae246d1b1aed","Type":"ContainerDied","Data":"ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf"} Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.585403 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32409013-6281-4048-8606-ae246d1b1aed","Type":"ContainerDied","Data":"40d83e7741eae69f4c0f7122c7d19f8db18c60df20ef5a776a31e5fff2316a2f"} Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.585424 4990 scope.go:117] "RemoveContainer" containerID="ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.585608 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.619088 4990 scope.go:117] "RemoveContainer" containerID="ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf" Dec 03 13:05:07 crc kubenswrapper[4990]: E1203 13:05:07.619784 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf\": container with ID starting with ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf not found: ID does not exist" containerID="ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.619821 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf"} err="failed to get container status \"ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf\": rpc error: code = NotFound desc = could not find container \"ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf\": container with ID starting with ad550e047f97fd2253b55a1685c6edbeea47ba5b622a44dbe32b313dd0e4cdbf not found: ID does not exist" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.641082 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.659645 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.678436 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.679742 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.688069 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.694669 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.827334 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e634575-20df-49bc-b2af-5322b408b702-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.827433 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e634575-20df-49bc-b2af-5322b408b702-config-data\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.827518 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgmhw\" (UniqueName: \"kubernetes.io/projected/9e634575-20df-49bc-b2af-5322b408b702-kube-api-access-fgmhw\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.876194 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 13:05:07 crc kubenswrapper[4990]: W1203 13:05:07.881036 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc704a76d_fc53_4741_9bd5_b7893e1c96a5.slice/crio-0c20eefce506b7c90cc7245ddb08e9f78a6939d65699c8d7ffc61c73864888dc WatchSource:0}: Error finding container 0c20eefce506b7c90cc7245ddb08e9f78a6939d65699c8d7ffc61c73864888dc: Status 404 returned error can't find the container with id 0c20eefce506b7c90cc7245ddb08e9f78a6939d65699c8d7ffc61c73864888dc Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.930491 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgmhw\" (UniqueName: \"kubernetes.io/projected/9e634575-20df-49bc-b2af-5322b408b702-kube-api-access-fgmhw\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.930683 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e634575-20df-49bc-b2af-5322b408b702-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.930758 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e634575-20df-49bc-b2af-5322b408b702-config-data\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.935155 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e634575-20df-49bc-b2af-5322b408b702-config-data\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.936275 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e634575-20df-49bc-b2af-5322b408b702-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:07 crc kubenswrapper[4990]: I1203 13:05:07.946056 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgmhw\" (UniqueName: \"kubernetes.io/projected/9e634575-20df-49bc-b2af-5322b408b702-kube-api-access-fgmhw\") pod \"nova-scheduler-0\" (UID: \"9e634575-20df-49bc-b2af-5322b408b702\") " pod="openstack/nova-scheduler-0" Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.001889 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.276070 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d459c9c-7ea1-42c1-996f-85f099f16e90" path="/var/lib/kubelet/pods/0d459c9c-7ea1-42c1-996f-85f099f16e90/volumes" Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.277181 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32409013-6281-4048-8606-ae246d1b1aed" path="/var/lib/kubelet/pods/32409013-6281-4048-8606-ae246d1b1aed/volumes" Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.459705 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 13:05:08 crc kubenswrapper[4990]: W1203 13:05:08.465844 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e634575_20df_49bc_b2af_5322b408b702.slice/crio-4b36ee2db7d8baebd4de3d0ef8b73c042f3bc96ec7d6049f6e1e14d186143fbf WatchSource:0}: Error finding container 4b36ee2db7d8baebd4de3d0ef8b73c042f3bc96ec7d6049f6e1e14d186143fbf: Status 404 returned error can't find the container with id 4b36ee2db7d8baebd4de3d0ef8b73c042f3bc96ec7d6049f6e1e14d186143fbf Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.598225 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c704a76d-fc53-4741-9bd5-b7893e1c96a5","Type":"ContainerStarted","Data":"dcc86fac24fa6d26243e4ff9b3ae964e8e996fd93acdc693db1a15d3f8e2a08d"} Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.598279 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c704a76d-fc53-4741-9bd5-b7893e1c96a5","Type":"ContainerStarted","Data":"0c20eefce506b7c90cc7245ddb08e9f78a6939d65699c8d7ffc61c73864888dc"} Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.600857 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e634575-20df-49bc-b2af-5322b408b702","Type":"ContainerStarted","Data":"4b36ee2db7d8baebd4de3d0ef8b73c042f3bc96ec7d6049f6e1e14d186143fbf"} Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.937780 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:34164->10.217.0.193:8775: read: connection reset by peer" Dec 03 13:05:08 crc kubenswrapper[4990]: I1203 13:05:08.938072 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:34160->10.217.0.193:8775: read: connection reset by peer" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.428083 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.550887 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-59vqb"] Dec 03 13:05:09 crc kubenswrapper[4990]: E1203 13:05:09.551298 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-metadata" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.551315 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-metadata" Dec 03 13:05:09 crc kubenswrapper[4990]: E1203 13:05:09.551352 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-log" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.551359 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-log" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.551552 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-metadata" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.551564 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerName="nova-metadata-log" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.553216 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.560822 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-combined-ca-bundle\") pod \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.560872 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d64wd\" (UniqueName: \"kubernetes.io/projected/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-kube-api-access-d64wd\") pod \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.560949 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-config-data\") pod \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.561039 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-nova-metadata-tls-certs\") pod \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.561097 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-logs\") pod \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\" (UID: \"0780aac2-023d-4f4a-8fef-ccbaf8af60c7\") " Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.562136 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-logs" (OuterVolumeSpecName: "logs") pod "0780aac2-023d-4f4a-8fef-ccbaf8af60c7" (UID: "0780aac2-023d-4f4a-8fef-ccbaf8af60c7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.572497 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-59vqb"] Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.576750 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-kube-api-access-d64wd" (OuterVolumeSpecName: "kube-api-access-d64wd") pod "0780aac2-023d-4f4a-8fef-ccbaf8af60c7" (UID: "0780aac2-023d-4f4a-8fef-ccbaf8af60c7"). InnerVolumeSpecName "kube-api-access-d64wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.613212 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0780aac2-023d-4f4a-8fef-ccbaf8af60c7" (UID: "0780aac2-023d-4f4a-8fef-ccbaf8af60c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.617143 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c704a76d-fc53-4741-9bd5-b7893e1c96a5","Type":"ContainerStarted","Data":"950b67e4b932b8a28de1c5a2d914b6f7b4292a39d4aa0e2a1ad2e0f1f3fddd51"} Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.619542 4990 generic.go:334] "Generic (PLEG): container finished" podID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" containerID="bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e" exitCode=0 Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.619655 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0780aac2-023d-4f4a-8fef-ccbaf8af60c7","Type":"ContainerDied","Data":"bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e"} Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.619728 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0780aac2-023d-4f4a-8fef-ccbaf8af60c7","Type":"ContainerDied","Data":"f610017222edeb5eb8ef3b109e1b2227e621b1457b053c543008297927963ca1"} Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.619828 4990 scope.go:117] "RemoveContainer" containerID="bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.620020 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.626056 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e634575-20df-49bc-b2af-5322b408b702","Type":"ContainerStarted","Data":"26b6eaacd51c9aea55fe7840cbb43bb0ab7910f9d1b358fbefa703924fed4ad1"} Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.627990 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-config-data" (OuterVolumeSpecName: "config-data") pod "0780aac2-023d-4f4a-8fef-ccbaf8af60c7" (UID: "0780aac2-023d-4f4a-8fef-ccbaf8af60c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.635182 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.63516131 podStartE2EDuration="2.63516131s" podCreationTimestamp="2025-12-03 13:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:05:09.630867786 +0000 UTC m=+1657.772779015" watchObservedRunningTime="2025-12-03 13:05:09.63516131 +0000 UTC m=+1657.777072539" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.661478 4990 scope.go:117] "RemoveContainer" containerID="df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.662893 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-utilities\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.663102 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-catalog-content\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.663329 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5cml\" (UniqueName: \"kubernetes.io/projected/2ee1ab71-a068-4971-bdca-b109a3952e1d-kube-api-access-g5cml\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.663446 4990 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.663484 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.663501 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d64wd\" (UniqueName: \"kubernetes.io/projected/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-kube-api-access-d64wd\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.663513 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.667432 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.667413055 podStartE2EDuration="2.667413055s" podCreationTimestamp="2025-12-03 13:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:05:09.654650737 +0000 UTC m=+1657.796561966" watchObservedRunningTime="2025-12-03 13:05:09.667413055 +0000 UTC m=+1657.809324284" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.685171 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0780aac2-023d-4f4a-8fef-ccbaf8af60c7" (UID: "0780aac2-023d-4f4a-8fef-ccbaf8af60c7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.685965 4990 scope.go:117] "RemoveContainer" containerID="bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e" Dec 03 13:05:09 crc kubenswrapper[4990]: E1203 13:05:09.686626 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e\": container with ID starting with bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e not found: ID does not exist" containerID="bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.686668 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e"} err="failed to get container status \"bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e\": rpc error: code = NotFound desc = could not find container \"bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e\": container with ID starting with bf7a2feac8b7bef9f42898afc637d9d9624c50a3e3df45fbfea6dd05ca45bb4e not found: ID does not exist" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.686695 4990 scope.go:117] "RemoveContainer" containerID="df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146" Dec 03 13:05:09 crc kubenswrapper[4990]: E1203 13:05:09.686996 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146\": container with ID starting with df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146 not found: ID does not exist" containerID="df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.687017 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146"} err="failed to get container status \"df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146\": rpc error: code = NotFound desc = could not find container \"df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146\": container with ID starting with df46723070126afc09d05b1f4a8e9d96bd2e87975bf4c1b72cbaf307b3c5f146 not found: ID does not exist" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.765601 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-catalog-content\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.765744 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5cml\" (UniqueName: \"kubernetes.io/projected/2ee1ab71-a068-4971-bdca-b109a3952e1d-kube-api-access-g5cml\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.765822 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-utilities\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.765903 4990 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0780aac2-023d-4f4a-8fef-ccbaf8af60c7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.766404 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-utilities\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.766531 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-catalog-content\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.781953 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5cml\" (UniqueName: \"kubernetes.io/projected/2ee1ab71-a068-4971-bdca-b109a3952e1d-kube-api-access-g5cml\") pod \"certified-operators-59vqb\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:09 crc kubenswrapper[4990]: I1203 13:05:09.917778 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.002880 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.019976 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.029572 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.031333 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.036281 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.036473 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.039346 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.179999 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.180063 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4737b59-9b3b-4bd5-8220-1346a05c998f-logs\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.180201 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-config-data\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.180227 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:10 crc kubenswrapper[4990]: I1203 13:05:10.180279 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkc2l\" (UniqueName: \"kubernetes.io/projected/f4737b59-9b3b-4bd5-8220-1346a05c998f-kube-api-access-xkc2l\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.283309 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-config-data\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.283343 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.283363 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkc2l\" (UniqueName: \"kubernetes.io/projected/f4737b59-9b3b-4bd5-8220-1346a05c998f-kube-api-access-xkc2l\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.283468 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.283489 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4737b59-9b3b-4bd5-8220-1346a05c998f-logs\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.283868 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4737b59-9b3b-4bd5-8220-1346a05c998f-logs\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.288419 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.292305 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.296830 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0780aac2-023d-4f4a-8fef-ccbaf8af60c7" path="/var/lib/kubelet/pods/0780aac2-023d-4f4a-8fef-ccbaf8af60c7/volumes" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.298568 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4737b59-9b3b-4bd5-8220-1346a05c998f-config-data\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.328258 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkc2l\" (UniqueName: \"kubernetes.io/projected/f4737b59-9b3b-4bd5-8220-1346a05c998f-kube-api-access-xkc2l\") pod \"nova-metadata-0\" (UID: \"f4737b59-9b3b-4bd5-8220-1346a05c998f\") " pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.380846 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.419137 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-59vqb"] Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.637729 4990 generic.go:334] "Generic (PLEG): container finished" podID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerID="0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda" exitCode=0 Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.637827 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerDied","Data":"0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda"} Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:10.638124 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerStarted","Data":"2405ee9e5e2539da413e3191e294bc5547a9f84a1e571e511a364598c0e3a4a5"} Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:11.656705 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerStarted","Data":"b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e"} Dec 03 13:05:11 crc kubenswrapper[4990]: I1203 13:05:11.928732 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 13:05:11 crc kubenswrapper[4990]: W1203 13:05:11.930642 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4737b59_9b3b_4bd5_8220_1346a05c998f.slice/crio-a1753331c6026b92bd447a39a0add14e93c673103ad0b652f377765d37bdb9f1 WatchSource:0}: Error finding container a1753331c6026b92bd447a39a0add14e93c673103ad0b652f377765d37bdb9f1: Status 404 returned error can't find the container with id a1753331c6026b92bd447a39a0add14e93c673103ad0b652f377765d37bdb9f1 Dec 03 13:05:12 crc kubenswrapper[4990]: I1203 13:05:12.667608 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4737b59-9b3b-4bd5-8220-1346a05c998f","Type":"ContainerStarted","Data":"840d232e93ed5758862c5a2f165f82a0f5f23acb453bc4c0b6bb8b02fc2790f2"} Dec 03 13:05:12 crc kubenswrapper[4990]: I1203 13:05:12.667932 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4737b59-9b3b-4bd5-8220-1346a05c998f","Type":"ContainerStarted","Data":"2559373b015a12f9764043e12b662cdaab18c579ef57710e73224dacb7439480"} Dec 03 13:05:12 crc kubenswrapper[4990]: I1203 13:05:12.667945 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4737b59-9b3b-4bd5-8220-1346a05c998f","Type":"ContainerStarted","Data":"a1753331c6026b92bd447a39a0add14e93c673103ad0b652f377765d37bdb9f1"} Dec 03 13:05:12 crc kubenswrapper[4990]: I1203 13:05:12.670051 4990 generic.go:334] "Generic (PLEG): container finished" podID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerID="b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e" exitCode=0 Dec 03 13:05:12 crc kubenswrapper[4990]: I1203 13:05:12.670120 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerDied","Data":"b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e"} Dec 03 13:05:12 crc kubenswrapper[4990]: I1203 13:05:12.695851 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.695828859 podStartE2EDuration="3.695828859s" podCreationTimestamp="2025-12-03 13:05:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:05:12.689513531 +0000 UTC m=+1660.831424790" watchObservedRunningTime="2025-12-03 13:05:12.695828859 +0000 UTC m=+1660.837740078" Dec 03 13:05:13 crc kubenswrapper[4990]: I1203 13:05:13.002678 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 13:05:13 crc kubenswrapper[4990]: I1203 13:05:13.680930 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerStarted","Data":"26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427"} Dec 03 13:05:13 crc kubenswrapper[4990]: I1203 13:05:13.706424 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-59vqb" podStartSLOduration=2.12706204 podStartE2EDuration="4.706400722s" podCreationTimestamp="2025-12-03 13:05:09 +0000 UTC" firstStartedPulling="2025-12-03 13:05:10.639525749 +0000 UTC m=+1658.781436978" lastFinishedPulling="2025-12-03 13:05:13.218864431 +0000 UTC m=+1661.360775660" observedRunningTime="2025-12-03 13:05:13.696828089 +0000 UTC m=+1661.838739338" watchObservedRunningTime="2025-12-03 13:05:13.706400722 +0000 UTC m=+1661.848311951" Dec 03 13:05:15 crc kubenswrapper[4990]: I1203 13:05:15.381709 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:05:15 crc kubenswrapper[4990]: I1203 13:05:15.382090 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 13:05:17 crc kubenswrapper[4990]: I1203 13:05:17.406381 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:05:17 crc kubenswrapper[4990]: I1203 13:05:17.406707 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 13:05:18 crc kubenswrapper[4990]: I1203 13:05:18.003013 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 13:05:18 crc kubenswrapper[4990]: I1203 13:05:18.030172 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 13:05:18 crc kubenswrapper[4990]: I1203 13:05:18.267314 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:05:18 crc kubenswrapper[4990]: E1203 13:05:18.267614 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:05:18 crc kubenswrapper[4990]: I1203 13:05:18.418766 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c704a76d-fc53-4741-9bd5-b7893e1c96a5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:05:18 crc kubenswrapper[4990]: I1203 13:05:18.418777 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c704a76d-fc53-4741-9bd5-b7893e1c96a5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:05:18 crc kubenswrapper[4990]: I1203 13:05:18.755794 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 13:05:19 crc kubenswrapper[4990]: I1203 13:05:19.918886 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:19 crc kubenswrapper[4990]: I1203 13:05:19.919229 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:19 crc kubenswrapper[4990]: I1203 13:05:19.968839 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:20 crc kubenswrapper[4990]: I1203 13:05:20.381785 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:05:20 crc kubenswrapper[4990]: I1203 13:05:20.381824 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 13:05:20 crc kubenswrapper[4990]: I1203 13:05:20.782250 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:20 crc kubenswrapper[4990]: I1203 13:05:20.829021 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-59vqb"] Dec 03 13:05:21 crc kubenswrapper[4990]: I1203 13:05:21.394646 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f4737b59-9b3b-4bd5-8220-1346a05c998f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:05:21 crc kubenswrapper[4990]: I1203 13:05:21.394628 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f4737b59-9b3b-4bd5-8220-1346a05c998f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.624054 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hvrql"] Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.627999 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.640863 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvrql"] Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.744036 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-utilities\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.744343 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxhwf\" (UniqueName: \"kubernetes.io/projected/03c39727-9e65-4865-a776-9c724710b725-kube-api-access-rxhwf\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.744513 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-catalog-content\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.760937 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-59vqb" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="registry-server" containerID="cri-o://26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427" gracePeriod=2 Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.846049 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxhwf\" (UniqueName: \"kubernetes.io/projected/03c39727-9e65-4865-a776-9c724710b725-kube-api-access-rxhwf\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.846142 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-catalog-content\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.846217 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-utilities\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.846646 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-catalog-content\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.846665 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-utilities\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.874431 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxhwf\" (UniqueName: \"kubernetes.io/projected/03c39727-9e65-4865-a776-9c724710b725-kube-api-access-rxhwf\") pod \"redhat-marketplace-hvrql\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.963910 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:22 crc kubenswrapper[4990]: I1203 13:05:22.974291 4990 scope.go:117] "RemoveContainer" containerID="560c3677bbbafe6087addfdc243849349f0dedcf30e74d70b7c1ff66bc4250ca" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.018368 4990 scope.go:117] "RemoveContainer" containerID="a5006df1726239cff9c557bb720ae173afdfaaf47878e383e9b290df9aab61d4" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.225696 4990 scope.go:117] "RemoveContainer" containerID="4fd7235ef2951fa44c32f912cca0258424f0d4ec8bf7af115ed1fa97eecc03ba" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.293604 4990 scope.go:117] "RemoveContainer" containerID="76daed6ff0ab3a3936010c09a4a3f3f32db810f207331e602d475b18359cb909" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.649204 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvrql"] Dec 03 13:05:23 crc kubenswrapper[4990]: W1203 13:05:23.650309 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03c39727_9e65_4865_a776_9c724710b725.slice/crio-07a6bb805d5142f4c1ba5579f22873989a4186f614f25facfa4871a54fbbd4a6 WatchSource:0}: Error finding container 07a6bb805d5142f4c1ba5579f22873989a4186f614f25facfa4871a54fbbd4a6: Status 404 returned error can't find the container with id 07a6bb805d5142f4c1ba5579f22873989a4186f614f25facfa4871a54fbbd4a6 Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.653597 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.776783 4990 generic.go:334] "Generic (PLEG): container finished" podID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerID="26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427" exitCode=0 Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.777223 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerDied","Data":"26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427"} Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.777264 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59vqb" event={"ID":"2ee1ab71-a068-4971-bdca-b109a3952e1d","Type":"ContainerDied","Data":"2405ee9e5e2539da413e3191e294bc5547a9f84a1e571e511a364598c0e3a4a5"} Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.777282 4990 scope.go:117] "RemoveContainer" containerID="26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.777415 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59vqb" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.780218 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-utilities\") pod \"2ee1ab71-a068-4971-bdca-b109a3952e1d\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.780354 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-catalog-content\") pod \"2ee1ab71-a068-4971-bdca-b109a3952e1d\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.780693 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvrql" event={"ID":"03c39727-9e65-4865-a776-9c724710b725","Type":"ContainerStarted","Data":"07a6bb805d5142f4c1ba5579f22873989a4186f614f25facfa4871a54fbbd4a6"} Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.781008 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5cml\" (UniqueName: \"kubernetes.io/projected/2ee1ab71-a068-4971-bdca-b109a3952e1d-kube-api-access-g5cml\") pod \"2ee1ab71-a068-4971-bdca-b109a3952e1d\" (UID: \"2ee1ab71-a068-4971-bdca-b109a3952e1d\") " Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.781324 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-utilities" (OuterVolumeSpecName: "utilities") pod "2ee1ab71-a068-4971-bdca-b109a3952e1d" (UID: "2ee1ab71-a068-4971-bdca-b109a3952e1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.781603 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.788275 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee1ab71-a068-4971-bdca-b109a3952e1d-kube-api-access-g5cml" (OuterVolumeSpecName: "kube-api-access-g5cml") pod "2ee1ab71-a068-4971-bdca-b109a3952e1d" (UID: "2ee1ab71-a068-4971-bdca-b109a3952e1d"). InnerVolumeSpecName "kube-api-access-g5cml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.807268 4990 scope.go:117] "RemoveContainer" containerID="b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.843647 4990 scope.go:117] "RemoveContainer" containerID="0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.872506 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ee1ab71-a068-4971-bdca-b109a3952e1d" (UID: "2ee1ab71-a068-4971-bdca-b109a3952e1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.883986 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee1ab71-a068-4971-bdca-b109a3952e1d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.884033 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5cml\" (UniqueName: \"kubernetes.io/projected/2ee1ab71-a068-4971-bdca-b109a3952e1d-kube-api-access-g5cml\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.896940 4990 scope.go:117] "RemoveContainer" containerID="26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427" Dec 03 13:05:23 crc kubenswrapper[4990]: E1203 13:05:23.897562 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427\": container with ID starting with 26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427 not found: ID does not exist" containerID="26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.897600 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427"} err="failed to get container status \"26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427\": rpc error: code = NotFound desc = could not find container \"26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427\": container with ID starting with 26573540f521cbd14fc8bbcdd22b565125d0a8d70d41555269b888987c2ba427 not found: ID does not exist" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.897626 4990 scope.go:117] "RemoveContainer" containerID="b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e" Dec 03 13:05:23 crc kubenswrapper[4990]: E1203 13:05:23.898079 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e\": container with ID starting with b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e not found: ID does not exist" containerID="b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.898135 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e"} err="failed to get container status \"b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e\": rpc error: code = NotFound desc = could not find container \"b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e\": container with ID starting with b45dd5da36f7300987d6750e766ea0339ce69b65050078dcde7dc6b21d1cd31e not found: ID does not exist" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.898174 4990 scope.go:117] "RemoveContainer" containerID="0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda" Dec 03 13:05:23 crc kubenswrapper[4990]: E1203 13:05:23.899533 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda\": container with ID starting with 0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda not found: ID does not exist" containerID="0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda" Dec 03 13:05:23 crc kubenswrapper[4990]: I1203 13:05:23.899577 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda"} err="failed to get container status \"0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda\": rpc error: code = NotFound desc = could not find container \"0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda\": container with ID starting with 0f425e0879f4c2b6ef75e11f0d1fed97770aa282afd867ba1b7849c335c03eda not found: ID does not exist" Dec 03 13:05:24 crc kubenswrapper[4990]: I1203 13:05:24.162112 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-59vqb"] Dec 03 13:05:24 crc kubenswrapper[4990]: I1203 13:05:24.175138 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-59vqb"] Dec 03 13:05:24 crc kubenswrapper[4990]: I1203 13:05:24.276841 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" path="/var/lib/kubelet/pods/2ee1ab71-a068-4971-bdca-b109a3952e1d/volumes" Dec 03 13:05:24 crc kubenswrapper[4990]: I1203 13:05:24.792443 4990 generic.go:334] "Generic (PLEG): container finished" podID="03c39727-9e65-4865-a776-9c724710b725" containerID="546a1ef912dde06407c6d739cc39028d386381d501c94ce8033a9875c547fe3b" exitCode=0 Dec 03 13:05:24 crc kubenswrapper[4990]: I1203 13:05:24.792499 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvrql" event={"ID":"03c39727-9e65-4865-a776-9c724710b725","Type":"ContainerDied","Data":"546a1ef912dde06407c6d739cc39028d386381d501c94ce8033a9875c547fe3b"} Dec 03 13:05:26 crc kubenswrapper[4990]: I1203 13:05:26.813417 4990 generic.go:334] "Generic (PLEG): container finished" podID="03c39727-9e65-4865-a776-9c724710b725" containerID="d49e602e9395e82f917f91b5d9ab1872c584be6da93decd14e0683b3ec32ec7f" exitCode=0 Dec 03 13:05:26 crc kubenswrapper[4990]: I1203 13:05:26.813528 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvrql" event={"ID":"03c39727-9e65-4865-a776-9c724710b725","Type":"ContainerDied","Data":"d49e602e9395e82f917f91b5d9ab1872c584be6da93decd14e0683b3ec32ec7f"} Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.132851 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.415304 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.416640 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.420912 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.426286 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.824488 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvrql" event={"ID":"03c39727-9e65-4865-a776-9c724710b725","Type":"ContainerStarted","Data":"dac06d20134555dcc54ebef75bc8e29100ebcf015c133b3803b6f4955e61c7df"} Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.824793 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.832474 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 13:05:27 crc kubenswrapper[4990]: I1203 13:05:27.848790 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hvrql" podStartSLOduration=3.256661806 podStartE2EDuration="5.848767307s" podCreationTimestamp="2025-12-03 13:05:22 +0000 UTC" firstStartedPulling="2025-12-03 13:05:24.79462338 +0000 UTC m=+1672.936534609" lastFinishedPulling="2025-12-03 13:05:27.386728881 +0000 UTC m=+1675.528640110" observedRunningTime="2025-12-03 13:05:27.842065219 +0000 UTC m=+1675.983976468" watchObservedRunningTime="2025-12-03 13:05:27.848767307 +0000 UTC m=+1675.990678546" Dec 03 13:05:30 crc kubenswrapper[4990]: I1203 13:05:30.387139 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 13:05:30 crc kubenswrapper[4990]: I1203 13:05:30.393384 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 13:05:30 crc kubenswrapper[4990]: I1203 13:05:30.393504 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 13:05:30 crc kubenswrapper[4990]: I1203 13:05:30.863395 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 13:05:32 crc kubenswrapper[4990]: I1203 13:05:32.271568 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:05:32 crc kubenswrapper[4990]: E1203 13:05:32.272106 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:05:32 crc kubenswrapper[4990]: I1203 13:05:32.964296 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:32 crc kubenswrapper[4990]: I1203 13:05:32.964385 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:33 crc kubenswrapper[4990]: I1203 13:05:33.013584 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:33 crc kubenswrapper[4990]: I1203 13:05:33.969786 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:34 crc kubenswrapper[4990]: I1203 13:05:34.049931 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvrql"] Dec 03 13:05:35 crc kubenswrapper[4990]: I1203 13:05:35.925727 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hvrql" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="registry-server" containerID="cri-o://dac06d20134555dcc54ebef75bc8e29100ebcf015c133b3803b6f4955e61c7df" gracePeriod=2 Dec 03 13:05:36 crc kubenswrapper[4990]: I1203 13:05:36.936723 4990 generic.go:334] "Generic (PLEG): container finished" podID="03c39727-9e65-4865-a776-9c724710b725" containerID="dac06d20134555dcc54ebef75bc8e29100ebcf015c133b3803b6f4955e61c7df" exitCode=0 Dec 03 13:05:36 crc kubenswrapper[4990]: I1203 13:05:36.936798 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvrql" event={"ID":"03c39727-9e65-4865-a776-9c724710b725","Type":"ContainerDied","Data":"dac06d20134555dcc54ebef75bc8e29100ebcf015c133b3803b6f4955e61c7df"} Dec 03 13:05:36 crc kubenswrapper[4990]: I1203 13:05:36.937076 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvrql" event={"ID":"03c39727-9e65-4865-a776-9c724710b725","Type":"ContainerDied","Data":"07a6bb805d5142f4c1ba5579f22873989a4186f614f25facfa4871a54fbbd4a6"} Dec 03 13:05:36 crc kubenswrapper[4990]: I1203 13:05:36.937095 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07a6bb805d5142f4c1ba5579f22873989a4186f614f25facfa4871a54fbbd4a6" Dec 03 13:05:36 crc kubenswrapper[4990]: I1203 13:05:36.975724 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.172187 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxhwf\" (UniqueName: \"kubernetes.io/projected/03c39727-9e65-4865-a776-9c724710b725-kube-api-access-rxhwf\") pod \"03c39727-9e65-4865-a776-9c724710b725\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.172252 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-catalog-content\") pod \"03c39727-9e65-4865-a776-9c724710b725\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.172555 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-utilities\") pod \"03c39727-9e65-4865-a776-9c724710b725\" (UID: \"03c39727-9e65-4865-a776-9c724710b725\") " Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.173771 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-utilities" (OuterVolumeSpecName: "utilities") pod "03c39727-9e65-4865-a776-9c724710b725" (UID: "03c39727-9e65-4865-a776-9c724710b725"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.197364 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c39727-9e65-4865-a776-9c724710b725-kube-api-access-rxhwf" (OuterVolumeSpecName: "kube-api-access-rxhwf") pod "03c39727-9e65-4865-a776-9c724710b725" (UID: "03c39727-9e65-4865-a776-9c724710b725"). InnerVolumeSpecName "kube-api-access-rxhwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.203839 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03c39727-9e65-4865-a776-9c724710b725" (UID: "03c39727-9e65-4865-a776-9c724710b725"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.275206 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxhwf\" (UniqueName: \"kubernetes.io/projected/03c39727-9e65-4865-a776-9c724710b725-kube-api-access-rxhwf\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.275239 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.275249 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c39727-9e65-4865-a776-9c724710b725-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.946544 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvrql" Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.989513 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvrql"] Dec 03 13:05:37 crc kubenswrapper[4990]: I1203 13:05:37.999386 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvrql"] Dec 03 13:05:38 crc kubenswrapper[4990]: I1203 13:05:38.274243 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c39727-9e65-4865-a776-9c724710b725" path="/var/lib/kubelet/pods/03c39727-9e65-4865-a776-9c724710b725/volumes" Dec 03 13:05:39 crc kubenswrapper[4990]: I1203 13:05:39.667945 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:05:40 crc kubenswrapper[4990]: I1203 13:05:40.567063 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:05:44 crc kubenswrapper[4990]: I1203 13:05:44.988891 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" containerID="cri-o://b6657e9f631e4d45ce3e912956fa8ee4bfd97492332688531321b57488cbbc45" gracePeriod=604796 Dec 03 13:05:45 crc kubenswrapper[4990]: I1203 13:05:45.040921 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" containerID="cri-o://b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413" gracePeriod=604795 Dec 03 13:05:45 crc kubenswrapper[4990]: I1203 13:05:45.264801 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:05:45 crc kubenswrapper[4990]: E1203 13:05:45.264996 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:05:45 crc kubenswrapper[4990]: I1203 13:05:45.385805 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 03 13:05:45 crc kubenswrapper[4990]: I1203 13:05:45.749362 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.265542 4990 generic.go:334] "Generic (PLEG): container finished" podID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerID="b6657e9f631e4d45ce3e912956fa8ee4bfd97492332688531321b57488cbbc45" exitCode=0 Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.265652 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d0a6648-1fc6-4928-ab9f-19829eaa0726","Type":"ContainerDied","Data":"b6657e9f631e4d45ce3e912956fa8ee4bfd97492332688531321b57488cbbc45"} Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.639680 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.644976 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.745913 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-tls\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.745966 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-erlang-cookie\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.745993 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-config-data\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746042 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c414845-915f-47b6-8a52-0e6f4e9ad886-erlang-cookie-secret\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746065 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-config-data\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746102 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-tls\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746139 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-confd\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746179 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf6sm\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-kube-api-access-bf6sm\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746197 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-confd\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746213 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d0a6648-1fc6-4928-ab9f-19829eaa0726-erlang-cookie-secret\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746262 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-erlang-cookie\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746284 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-plugins\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746300 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746327 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-server-conf\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746377 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d0a6648-1fc6-4928-ab9f-19829eaa0726-pod-info\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746395 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c414845-915f-47b6-8a52-0e6f4e9ad886-pod-info\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746416 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnjz5\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-kube-api-access-xnjz5\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746441 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-plugins-conf\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746487 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746518 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-server-conf\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746556 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-plugins-conf\") pod \"5c414845-915f-47b6-8a52-0e6f4e9ad886\" (UID: \"5c414845-915f-47b6-8a52-0e6f4e9ad886\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.746583 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-plugins\") pod \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\" (UID: \"4d0a6648-1fc6-4928-ab9f-19829eaa0726\") " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.747259 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.747283 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.748864 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.754515 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.754819 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-kube-api-access-xnjz5" (OuterVolumeSpecName: "kube-api-access-xnjz5") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "kube-api-access-xnjz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.754884 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c414845-915f-47b6-8a52-0e6f4e9ad886-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.755585 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4d0a6648-1fc6-4928-ab9f-19829eaa0726-pod-info" (OuterVolumeSpecName: "pod-info") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.755677 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.755908 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.755951 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.756184 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.759051 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-kube-api-access-bf6sm" (OuterVolumeSpecName: "kube-api-access-bf6sm") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "kube-api-access-bf6sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.760110 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.763990 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d0a6648-1fc6-4928-ab9f-19829eaa0726-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.768753 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.779284 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5c414845-915f-47b6-8a52-0e6f4e9ad886-pod-info" (OuterVolumeSpecName: "pod-info") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.781800 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-config-data" (OuterVolumeSpecName: "config-data") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.799698 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-config-data" (OuterVolumeSpecName: "config-data") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.825821 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-server-conf" (OuterVolumeSpecName: "server-conf") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.850135 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-server-conf" (OuterVolumeSpecName: "server-conf") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858640 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858672 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858684 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858693 4990 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5c414845-915f-47b6-8a52-0e6f4e9ad886-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858702 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858711 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858720 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf6sm\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-kube-api-access-bf6sm\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858728 4990 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d0a6648-1fc6-4928-ab9f-19829eaa0726-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858736 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858747 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858777 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858787 4990 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858797 4990 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d0a6648-1fc6-4928-ab9f-19829eaa0726-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858804 4990 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5c414845-915f-47b6-8a52-0e6f4e9ad886-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858813 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnjz5\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-kube-api-access-xnjz5\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858823 4990 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858837 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858845 4990 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d0a6648-1fc6-4928-ab9f-19829eaa0726-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858854 4990 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5c414845-915f-47b6-8a52-0e6f4e9ad886-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.858878 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.882638 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.888375 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.891132 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5c414845-915f-47b6-8a52-0e6f4e9ad886" (UID: "5c414845-915f-47b6-8a52-0e6f4e9ad886"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.919743 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4d0a6648-1fc6-4928-ab9f-19829eaa0726" (UID: "4d0a6648-1fc6-4928-ab9f-19829eaa0726"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.960201 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5c414845-915f-47b6-8a52-0e6f4e9ad886-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.960240 4990 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d0a6648-1fc6-4928-ab9f-19829eaa0726-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.960249 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:51 crc kubenswrapper[4990]: I1203 13:05:51.960261 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.279963 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d0a6648-1fc6-4928-ab9f-19829eaa0726","Type":"ContainerDied","Data":"4b38f92942fcc00996e1422c0650b0fe008bd05143650cb6c15aca4ce203b726"} Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.280017 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.280570 4990 scope.go:117] "RemoveContainer" containerID="b6657e9f631e4d45ce3e912956fa8ee4bfd97492332688531321b57488cbbc45" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.283537 4990 generic.go:334] "Generic (PLEG): container finished" podID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerID="b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413" exitCode=0 Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.283576 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5c414845-915f-47b6-8a52-0e6f4e9ad886","Type":"ContainerDied","Data":"b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413"} Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.283604 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5c414845-915f-47b6-8a52-0e6f4e9ad886","Type":"ContainerDied","Data":"c54fe6cf6c4228759e0ea76c010664c1df5fc18423828e31772c0fc231e1d5b8"} Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.283698 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.308849 4990 scope.go:117] "RemoveContainer" containerID="68b733f8c158a6602693114ab4de70f30583ca7579fb2a07014dc57a6407e951" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.341360 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.356062 4990 scope.go:117] "RemoveContainer" containerID="b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.358809 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.388142 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.403992 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.408384 4990 scope.go:117] "RemoveContainer" containerID="b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.419641 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420333 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420356 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420384 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="registry-server" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420392 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="registry-server" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420416 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="setup-container" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420427 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="setup-container" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420445 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="extract-content" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420475 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="extract-content" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420488 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420496 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420511 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="setup-container" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420519 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="setup-container" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420548 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="extract-content" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420555 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="extract-content" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420566 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="extract-utilities" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420575 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="extract-utilities" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420586 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="extract-utilities" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420595 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="extract-utilities" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.420621 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="registry-server" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420629 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="registry-server" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420886 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee1ab71-a068-4971-bdca-b109a3952e1d" containerName="registry-server" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420907 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c39727-9e65-4865-a776-9c724710b725" containerName="registry-server" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420922 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" containerName="rabbitmq" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.420933 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" containerName="rabbitmq" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.422524 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.431386 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.431683 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.431845 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.432075 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.432901 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-j442g" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.433102 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.433292 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.438881 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.459596 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.461853 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.466416 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.466646 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.466735 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.466805 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.469426 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.469719 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mtzzs" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.470035 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.473713 4990 scope.go:117] "RemoveContainer" containerID="b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.475980 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413\": container with ID starting with b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413 not found: ID does not exist" containerID="b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.476027 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413"} err="failed to get container status \"b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413\": rpc error: code = NotFound desc = could not find container \"b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413\": container with ID starting with b8d3b497cd9cb9c01b34fb32847a32386af44fd2b52aa424ff3c03c24a44c413 not found: ID does not exist" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.476085 4990 scope.go:117] "RemoveContainer" containerID="b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1" Dec 03 13:05:52 crc kubenswrapper[4990]: E1203 13:05:52.476676 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1\": container with ID starting with b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1 not found: ID does not exist" containerID="b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.476720 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1"} err="failed to get container status \"b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1\": rpc error: code = NotFound desc = could not find container \"b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1\": container with ID starting with b19e557772f338285746c4be2c6bb6425a2e1959d550a72a7d10072ec444b8c1 not found: ID does not exist" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.483767 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.576654 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.576722 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.576751 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.576779 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.576813 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.576833 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577061 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/be4855bd-7c56-4d1c-af0d-4017083409a6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577108 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577148 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98e2d7f8-4691-44ca-a46f-9e08c82844a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577182 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98e2d7f8-4691-44ca-a46f-9e08c82844a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577213 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/be4855bd-7c56-4d1c-af0d-4017083409a6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577296 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577357 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577413 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577443 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577518 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577543 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g6f9\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-kube-api-access-9g6f9\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577568 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577588 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577620 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577649 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl7nj\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-kube-api-access-kl7nj\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.577709 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.679798 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.679861 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.679888 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.679910 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.679953 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.679972 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680028 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/be4855bd-7c56-4d1c-af0d-4017083409a6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680044 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680061 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98e2d7f8-4691-44ca-a46f-9e08c82844a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680078 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98e2d7f8-4691-44ca-a46f-9e08c82844a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680112 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/be4855bd-7c56-4d1c-af0d-4017083409a6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680142 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680184 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680208 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680228 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680271 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680276 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680838 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680923 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681124 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.680289 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g6f9\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-kube-api-access-9g6f9\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681235 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681262 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681289 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681316 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl7nj\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-kube-api-access-kl7nj\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681354 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681440 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681471 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681523 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.681749 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.682253 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.682914 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.683065 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98e2d7f8-4691-44ca-a46f-9e08c82844a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.683706 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/be4855bd-7c56-4d1c-af0d-4017083409a6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.692632 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98e2d7f8-4691-44ca-a46f-9e08c82844a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.692662 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/be4855bd-7c56-4d1c-af0d-4017083409a6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.692869 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.693085 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98e2d7f8-4691-44ca-a46f-9e08c82844a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.693351 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.693284 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/be4855bd-7c56-4d1c-af0d-4017083409a6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.696465 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.720934 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g6f9\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-kube-api-access-9g6f9\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.722627 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl7nj\" (UniqueName: \"kubernetes.io/projected/be4855bd-7c56-4d1c-af0d-4017083409a6-kube-api-access-kl7nj\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.722874 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98e2d7f8-4691-44ca-a46f-9e08c82844a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.728552 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"be4855bd-7c56-4d1c-af0d-4017083409a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.738050 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"98e2d7f8-4691-44ca-a46f-9e08c82844a2\") " pod="openstack/rabbitmq-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.799760 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:05:52 crc kubenswrapper[4990]: I1203 13:05:52.811674 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 13:05:53 crc kubenswrapper[4990]: I1203 13:05:53.343242 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 13:05:53 crc kubenswrapper[4990]: W1203 13:05:53.418765 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe4855bd_7c56_4d1c_af0d_4017083409a6.slice/crio-d855c37d36d0ece46ea6bfcd1f60a58731dfc2950ce25a5476a0fdadf76c8d3d WatchSource:0}: Error finding container d855c37d36d0ece46ea6bfcd1f60a58731dfc2950ce25a5476a0fdadf76c8d3d: Status 404 returned error can't find the container with id d855c37d36d0ece46ea6bfcd1f60a58731dfc2950ce25a5476a0fdadf76c8d3d Dec 03 13:05:53 crc kubenswrapper[4990]: I1203 13:05:53.423874 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 13:05:54 crc kubenswrapper[4990]: I1203 13:05:54.276695 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d0a6648-1fc6-4928-ab9f-19829eaa0726" path="/var/lib/kubelet/pods/4d0a6648-1fc6-4928-ab9f-19829eaa0726/volumes" Dec 03 13:05:54 crc kubenswrapper[4990]: I1203 13:05:54.278540 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c414845-915f-47b6-8a52-0e6f4e9ad886" path="/var/lib/kubelet/pods/5c414845-915f-47b6-8a52-0e6f4e9ad886/volumes" Dec 03 13:05:54 crc kubenswrapper[4990]: I1203 13:05:54.305270 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"be4855bd-7c56-4d1c-af0d-4017083409a6","Type":"ContainerStarted","Data":"d855c37d36d0ece46ea6bfcd1f60a58731dfc2950ce25a5476a0fdadf76c8d3d"} Dec 03 13:05:54 crc kubenswrapper[4990]: I1203 13:05:54.306822 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98e2d7f8-4691-44ca-a46f-9e08c82844a2","Type":"ContainerStarted","Data":"d53a0907a0b8413d3466778bc675b0c3f06637259a3cb726ef3bd27a31e256ea"} Dec 03 13:05:55 crc kubenswrapper[4990]: I1203 13:05:55.321197 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"be4855bd-7c56-4d1c-af0d-4017083409a6","Type":"ContainerStarted","Data":"ebe41216183de02cbcd1670bf49b62762c56f27fa2e2e9b0e98936bae783e306"} Dec 03 13:05:56 crc kubenswrapper[4990]: I1203 13:05:56.348988 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98e2d7f8-4691-44ca-a46f-9e08c82844a2","Type":"ContainerStarted","Data":"333b1b150d33a20beee2ff3318f0610447871eb6759399bd2c8959e1a749cd0c"} Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.225175 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2t8z5"] Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.227134 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.229585 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.235880 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2t8z5"] Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.265682 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:05:57 crc kubenswrapper[4990]: E1203 13:05:57.265889 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.278827 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-svc\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.278881 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.278931 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.278991 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-config\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.279073 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.279097 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.279118 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgmp7\" (UniqueName: \"kubernetes.io/projected/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-kube-api-access-wgmp7\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.380839 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.380944 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-config\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.381051 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.381071 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.381104 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgmp7\" (UniqueName: \"kubernetes.io/projected/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-kube-api-access-wgmp7\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.381200 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-svc\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.381243 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.382939 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.384022 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-config\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.384612 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-svc\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.384770 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.384803 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.385192 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.411654 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgmp7\" (UniqueName: \"kubernetes.io/projected/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-kube-api-access-wgmp7\") pod \"dnsmasq-dns-5576978c7c-2t8z5\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:57 crc kubenswrapper[4990]: I1203 13:05:57.545428 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:58 crc kubenswrapper[4990]: I1203 13:05:58.025801 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2t8z5"] Dec 03 13:05:58 crc kubenswrapper[4990]: W1203 13:05:58.037898 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0de0959_b51c_4a8a_b6ae_9154d4cf0e6b.slice/crio-2976bc2c6e2fa8fad07cb24b9b4c9a050f4833b77340987a23f195ec40af21b8 WatchSource:0}: Error finding container 2976bc2c6e2fa8fad07cb24b9b4c9a050f4833b77340987a23f195ec40af21b8: Status 404 returned error can't find the container with id 2976bc2c6e2fa8fad07cb24b9b4c9a050f4833b77340987a23f195ec40af21b8 Dec 03 13:05:58 crc kubenswrapper[4990]: I1203 13:05:58.372008 4990 generic.go:334] "Generic (PLEG): container finished" podID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerID="d483f160eba871c41b3d54d0b38f5a51e4019a6933d29a25746abc6ad21ef776" exitCode=0 Dec 03 13:05:58 crc kubenswrapper[4990]: I1203 13:05:58.372148 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" event={"ID":"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b","Type":"ContainerDied","Data":"d483f160eba871c41b3d54d0b38f5a51e4019a6933d29a25746abc6ad21ef776"} Dec 03 13:05:58 crc kubenswrapper[4990]: I1203 13:05:58.372634 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" event={"ID":"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b","Type":"ContainerStarted","Data":"2976bc2c6e2fa8fad07cb24b9b4c9a050f4833b77340987a23f195ec40af21b8"} Dec 03 13:05:59 crc kubenswrapper[4990]: I1203 13:05:59.384938 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" event={"ID":"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b","Type":"ContainerStarted","Data":"71a396acd2bde86b0cd63d9828efdd30aaeb6d973796d626497704ab82569686"} Dec 03 13:05:59 crc kubenswrapper[4990]: I1203 13:05:59.385475 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:05:59 crc kubenswrapper[4990]: I1203 13:05:59.405290 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" podStartSLOduration=2.405267929 podStartE2EDuration="2.405267929s" podCreationTimestamp="2025-12-03 13:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:05:59.40488901 +0000 UTC m=+1707.546800239" watchObservedRunningTime="2025-12-03 13:05:59.405267929 +0000 UTC m=+1707.547179158" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.547764 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.627918 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-b24tp"] Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.628385 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="dnsmasq-dns" containerID="cri-o://1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023" gracePeriod=10 Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.807354 4990 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.200:5353: connect: connection refused" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.883534 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-s9vgm"] Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.885821 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.900988 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-s9vgm"] Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.906887 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spkrb\" (UniqueName: \"kubernetes.io/projected/f65f22f9-af79-4a63-bfc8-590d88507cb3-kube-api-access-spkrb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.907006 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.907960 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.907997 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.908080 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.908125 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-config\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:07 crc kubenswrapper[4990]: I1203 13:06:07.908145 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010305 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spkrb\" (UniqueName: \"kubernetes.io/projected/f65f22f9-af79-4a63-bfc8-590d88507cb3-kube-api-access-spkrb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010398 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010441 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010487 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010548 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010589 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-config\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.010613 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.011791 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-config\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.011846 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.011924 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.011981 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.012024 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.012660 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65f22f9-af79-4a63-bfc8-590d88507cb3-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.037719 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spkrb\" (UniqueName: \"kubernetes.io/projected/f65f22f9-af79-4a63-bfc8-590d88507cb3-kube-api-access-spkrb\") pod \"dnsmasq-dns-8c6f6df99-s9vgm\" (UID: \"f65f22f9-af79-4a63-bfc8-590d88507cb3\") " pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.138259 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.213367 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-nb\") pod \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.213428 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qcg4\" (UniqueName: \"kubernetes.io/projected/32c217f2-3a1b-47f5-91d9-b6634935c9f0-kube-api-access-7qcg4\") pod \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.213592 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-swift-storage-0\") pod \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.213672 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-config\") pod \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.213693 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-svc\") pod \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.213759 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-sb\") pod \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\" (UID: \"32c217f2-3a1b-47f5-91d9-b6634935c9f0\") " Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.214569 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.235598 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c217f2-3a1b-47f5-91d9-b6634935c9f0-kube-api-access-7qcg4" (OuterVolumeSpecName: "kube-api-access-7qcg4") pod "32c217f2-3a1b-47f5-91d9-b6634935c9f0" (UID: "32c217f2-3a1b-47f5-91d9-b6634935c9f0"). InnerVolumeSpecName "kube-api-access-7qcg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.268551 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32c217f2-3a1b-47f5-91d9-b6634935c9f0" (UID: "32c217f2-3a1b-47f5-91d9-b6634935c9f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.271686 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "32c217f2-3a1b-47f5-91d9-b6634935c9f0" (UID: "32c217f2-3a1b-47f5-91d9-b6634935c9f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.278328 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "32c217f2-3a1b-47f5-91d9-b6634935c9f0" (UID: "32c217f2-3a1b-47f5-91d9-b6634935c9f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.279614 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-config" (OuterVolumeSpecName: "config") pod "32c217f2-3a1b-47f5-91d9-b6634935c9f0" (UID: "32c217f2-3a1b-47f5-91d9-b6634935c9f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.282060 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "32c217f2-3a1b-47f5-91d9-b6634935c9f0" (UID: "32c217f2-3a1b-47f5-91d9-b6634935c9f0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.315486 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.315514 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.315524 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qcg4\" (UniqueName: \"kubernetes.io/projected/32c217f2-3a1b-47f5-91d9-b6634935c9f0-kube-api-access-7qcg4\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.315535 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.315545 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.315557 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32c217f2-3a1b-47f5-91d9-b6634935c9f0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.467051 4990 generic.go:334] "Generic (PLEG): container finished" podID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerID="1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023" exitCode=0 Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.467283 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" event={"ID":"32c217f2-3a1b-47f5-91d9-b6634935c9f0","Type":"ContainerDied","Data":"1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023"} Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.467387 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.467402 4990 scope.go:117] "RemoveContainer" containerID="1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.467387 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-b24tp" event={"ID":"32c217f2-3a1b-47f5-91d9-b6634935c9f0","Type":"ContainerDied","Data":"7f6ba03adb981fe163b06d93244df2469be04420e13132acee6fee0bc8012157"} Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.494358 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-b24tp"] Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.496761 4990 scope.go:117] "RemoveContainer" containerID="e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.509094 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-b24tp"] Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.529535 4990 scope.go:117] "RemoveContainer" containerID="1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023" Dec 03 13:06:08 crc kubenswrapper[4990]: E1203 13:06:08.530133 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023\": container with ID starting with 1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023 not found: ID does not exist" containerID="1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.530201 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023"} err="failed to get container status \"1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023\": rpc error: code = NotFound desc = could not find container \"1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023\": container with ID starting with 1d49872a8320bb82025dbafe2a62a3981599164267b8f210fa92c88a2ac9a023 not found: ID does not exist" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.530255 4990 scope.go:117] "RemoveContainer" containerID="e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230" Dec 03 13:06:08 crc kubenswrapper[4990]: E1203 13:06:08.531367 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230\": container with ID starting with e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230 not found: ID does not exist" containerID="e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.531417 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230"} err="failed to get container status \"e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230\": rpc error: code = NotFound desc = could not find container \"e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230\": container with ID starting with e639f87a2ce4ad9c9a02bbdea276a3629efaec70f735da56fef924a89901f230 not found: ID does not exist" Dec 03 13:06:08 crc kubenswrapper[4990]: I1203 13:06:08.679881 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-s9vgm"] Dec 03 13:06:08 crc kubenswrapper[4990]: W1203 13:06:08.682924 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf65f22f9_af79_4a63_bfc8_590d88507cb3.slice/crio-6d2f44642e69c68ce28ca71c1936d3fe3bde13ef001e12642bf2ccc73ea3e6e7 WatchSource:0}: Error finding container 6d2f44642e69c68ce28ca71c1936d3fe3bde13ef001e12642bf2ccc73ea3e6e7: Status 404 returned error can't find the container with id 6d2f44642e69c68ce28ca71c1936d3fe3bde13ef001e12642bf2ccc73ea3e6e7 Dec 03 13:06:09 crc kubenswrapper[4990]: I1203 13:06:09.477571 4990 generic.go:334] "Generic (PLEG): container finished" podID="f65f22f9-af79-4a63-bfc8-590d88507cb3" containerID="94bb4697f16719b63b3949abbcf71b23135822a3272690cdba9beecb71a05db4" exitCode=0 Dec 03 13:06:09 crc kubenswrapper[4990]: I1203 13:06:09.477888 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" event={"ID":"f65f22f9-af79-4a63-bfc8-590d88507cb3","Type":"ContainerDied","Data":"94bb4697f16719b63b3949abbcf71b23135822a3272690cdba9beecb71a05db4"} Dec 03 13:06:09 crc kubenswrapper[4990]: I1203 13:06:09.477963 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" event={"ID":"f65f22f9-af79-4a63-bfc8-590d88507cb3","Type":"ContainerStarted","Data":"6d2f44642e69c68ce28ca71c1936d3fe3bde13ef001e12642bf2ccc73ea3e6e7"} Dec 03 13:06:10 crc kubenswrapper[4990]: I1203 13:06:10.285678 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" path="/var/lib/kubelet/pods/32c217f2-3a1b-47f5-91d9-b6634935c9f0/volumes" Dec 03 13:06:10 crc kubenswrapper[4990]: I1203 13:06:10.491585 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" event={"ID":"f65f22f9-af79-4a63-bfc8-590d88507cb3","Type":"ContainerStarted","Data":"c80b7964317511b4590f9777530dd36d3c97a25a6ee7984dc8a5ea3b16ffe222"} Dec 03 13:06:10 crc kubenswrapper[4990]: I1203 13:06:10.491794 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:10 crc kubenswrapper[4990]: I1203 13:06:10.515164 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" podStartSLOduration=3.51515202 podStartE2EDuration="3.51515202s" podCreationTimestamp="2025-12-03 13:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:06:10.512226729 +0000 UTC m=+1718.654137958" watchObservedRunningTime="2025-12-03 13:06:10.51515202 +0000 UTC m=+1718.657063249" Dec 03 13:06:12 crc kubenswrapper[4990]: I1203 13:06:12.274695 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:06:12 crc kubenswrapper[4990]: E1203 13:06:12.275239 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.216823 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-s9vgm" Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.289085 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2t8z5"] Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.289327 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerName="dnsmasq-dns" containerID="cri-o://71a396acd2bde86b0cd63d9828efdd30aaeb6d973796d626497704ab82569686" gracePeriod=10 Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.597783 4990 generic.go:334] "Generic (PLEG): container finished" podID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerID="71a396acd2bde86b0cd63d9828efdd30aaeb6d973796d626497704ab82569686" exitCode=0 Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.599621 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" event={"ID":"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b","Type":"ContainerDied","Data":"71a396acd2bde86b0cd63d9828efdd30aaeb6d973796d626497704ab82569686"} Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.805722 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.942420 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-svc\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.942513 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-config\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.943335 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-swift-storage-0\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.943380 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-sb\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.943490 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-openstack-edpm-ipam\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.943520 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-nb\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.943597 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgmp7\" (UniqueName: \"kubernetes.io/projected/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-kube-api-access-wgmp7\") pod \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\" (UID: \"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b\") " Dec 03 13:06:18 crc kubenswrapper[4990]: I1203 13:06:18.959914 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-kube-api-access-wgmp7" (OuterVolumeSpecName: "kube-api-access-wgmp7") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "kube-api-access-wgmp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:18.998782 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-config" (OuterVolumeSpecName: "config") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:18.999695 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.003923 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.014068 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.022531 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.026861 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" (UID: "f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.045699 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.045949 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.046020 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgmp7\" (UniqueName: \"kubernetes.io/projected/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-kube-api-access-wgmp7\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.046082 4990 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.046154 4990 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.046211 4990 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.046261 4990 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.611469 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" event={"ID":"f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b","Type":"ContainerDied","Data":"2976bc2c6e2fa8fad07cb24b9b4c9a050f4833b77340987a23f195ec40af21b8"} Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.611828 4990 scope.go:117] "RemoveContainer" containerID="71a396acd2bde86b0cd63d9828efdd30aaeb6d973796d626497704ab82569686" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.611524 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2t8z5" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.673372 4990 scope.go:117] "RemoveContainer" containerID="d483f160eba871c41b3d54d0b38f5a51e4019a6933d29a25746abc6ad21ef776" Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.675878 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2t8z5"] Dec 03 13:06:19 crc kubenswrapper[4990]: I1203 13:06:19.684571 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2t8z5"] Dec 03 13:06:20 crc kubenswrapper[4990]: I1203 13:06:20.276112 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" path="/var/lib/kubelet/pods/f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b/volumes" Dec 03 13:06:24 crc kubenswrapper[4990]: I1203 13:06:24.266706 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:06:24 crc kubenswrapper[4990]: E1203 13:06:24.267522 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:06:29 crc kubenswrapper[4990]: I1203 13:06:29.755107 4990 generic.go:334] "Generic (PLEG): container finished" podID="be4855bd-7c56-4d1c-af0d-4017083409a6" containerID="ebe41216183de02cbcd1670bf49b62762c56f27fa2e2e9b0e98936bae783e306" exitCode=0 Dec 03 13:06:29 crc kubenswrapper[4990]: I1203 13:06:29.755197 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"be4855bd-7c56-4d1c-af0d-4017083409a6","Type":"ContainerDied","Data":"ebe41216183de02cbcd1670bf49b62762c56f27fa2e2e9b0e98936bae783e306"} Dec 03 13:06:29 crc kubenswrapper[4990]: I1203 13:06:29.762408 4990 generic.go:334] "Generic (PLEG): container finished" podID="98e2d7f8-4691-44ca-a46f-9e08c82844a2" containerID="333b1b150d33a20beee2ff3318f0610447871eb6759399bd2c8959e1a749cd0c" exitCode=0 Dec 03 13:06:29 crc kubenswrapper[4990]: I1203 13:06:29.762444 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98e2d7f8-4691-44ca-a46f-9e08c82844a2","Type":"ContainerDied","Data":"333b1b150d33a20beee2ff3318f0610447871eb6759399bd2c8959e1a749cd0c"} Dec 03 13:06:30 crc kubenswrapper[4990]: I1203 13:06:30.775499 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"be4855bd-7c56-4d1c-af0d-4017083409a6","Type":"ContainerStarted","Data":"d0ed4488f7fe0459f981051aee855c55de9fee5077b0cda1611f1b90f6a4731b"} Dec 03 13:06:30 crc kubenswrapper[4990]: I1203 13:06:30.776088 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:06:30 crc kubenswrapper[4990]: I1203 13:06:30.779666 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98e2d7f8-4691-44ca-a46f-9e08c82844a2","Type":"ContainerStarted","Data":"95f1f44a5cdbb349ba40053772c33af3537dfb64ff1aa8504c2b4eda5fab1784"} Dec 03 13:06:30 crc kubenswrapper[4990]: I1203 13:06:30.780173 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 13:06:30 crc kubenswrapper[4990]: I1203 13:06:30.805409 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.805393451 podStartE2EDuration="38.805393451s" podCreationTimestamp="2025-12-03 13:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:06:30.800488313 +0000 UTC m=+1738.942399562" watchObservedRunningTime="2025-12-03 13:06:30.805393451 +0000 UTC m=+1738.947304680" Dec 03 13:06:30 crc kubenswrapper[4990]: I1203 13:06:30.838193 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.838176228 podStartE2EDuration="38.838176228s" podCreationTimestamp="2025-12-03 13:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:06:30.826922538 +0000 UTC m=+1738.968833767" watchObservedRunningTime="2025-12-03 13:06:30.838176228 +0000 UTC m=+1738.980087457" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.226628 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z"] Dec 03 13:06:31 crc kubenswrapper[4990]: E1203 13:06:31.227087 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerName="dnsmasq-dns" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.227109 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerName="dnsmasq-dns" Dec 03 13:06:31 crc kubenswrapper[4990]: E1203 13:06:31.227136 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerName="init" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.227144 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerName="init" Dec 03 13:06:31 crc kubenswrapper[4990]: E1203 13:06:31.227163 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="dnsmasq-dns" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.227171 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="dnsmasq-dns" Dec 03 13:06:31 crc kubenswrapper[4990]: E1203 13:06:31.227193 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="init" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.227200 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="init" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.227405 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0de0959-b51c-4a8a-b6ae-9154d4cf0e6b" containerName="dnsmasq-dns" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.227438 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c217f2-3a1b-47f5-91d9-b6634935c9f0" containerName="dnsmasq-dns" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.230179 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.232574 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.232661 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.232585 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.236939 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.240638 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z"] Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.316898 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.316971 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvqsd\" (UniqueName: \"kubernetes.io/projected/3c11a50b-b2d3-45be-9999-bcd93b698b1c-kube-api-access-xvqsd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.317028 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.317264 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.420002 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.420097 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvqsd\" (UniqueName: \"kubernetes.io/projected/3c11a50b-b2d3-45be-9999-bcd93b698b1c-kube-api-access-xvqsd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.420150 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.420208 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.428553 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.428565 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.431061 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.444331 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvqsd\" (UniqueName: \"kubernetes.io/projected/3c11a50b-b2d3-45be-9999-bcd93b698b1c-kube-api-access-xvqsd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p972z\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:31 crc kubenswrapper[4990]: I1203 13:06:31.555720 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:06:32 crc kubenswrapper[4990]: I1203 13:06:32.166920 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z"] Dec 03 13:06:32 crc kubenswrapper[4990]: I1203 13:06:32.812598 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" event={"ID":"3c11a50b-b2d3-45be-9999-bcd93b698b1c","Type":"ContainerStarted","Data":"ce04577d5d0347a4fd9e0dcd7dbb4b8bb97003abc5a004b090a94284cc260b03"} Dec 03 13:06:38 crc kubenswrapper[4990]: I1203 13:06:38.269011 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:06:38 crc kubenswrapper[4990]: E1203 13:06:38.272222 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:06:41 crc kubenswrapper[4990]: I1203 13:06:41.626781 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:06:41 crc kubenswrapper[4990]: I1203 13:06:41.941926 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" event={"ID":"3c11a50b-b2d3-45be-9999-bcd93b698b1c","Type":"ContainerStarted","Data":"1e319ccf1dff713fbf85874fa9eb86be0a1576b306741a798fc6a6158f4f4748"} Dec 03 13:06:41 crc kubenswrapper[4990]: I1203 13:06:41.967314 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" podStartSLOduration=1.517222869 podStartE2EDuration="10.967293242s" podCreationTimestamp="2025-12-03 13:06:31 +0000 UTC" firstStartedPulling="2025-12-03 13:06:32.17390385 +0000 UTC m=+1740.315815079" lastFinishedPulling="2025-12-03 13:06:41.623974213 +0000 UTC m=+1749.765885452" observedRunningTime="2025-12-03 13:06:41.959862303 +0000 UTC m=+1750.101773542" watchObservedRunningTime="2025-12-03 13:06:41.967293242 +0000 UTC m=+1750.109204471" Dec 03 13:06:42 crc kubenswrapper[4990]: I1203 13:06:42.804694 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 13:06:42 crc kubenswrapper[4990]: I1203 13:06:42.815643 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 13:06:51 crc kubenswrapper[4990]: I1203 13:06:51.265411 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:06:51 crc kubenswrapper[4990]: E1203 13:06:51.266610 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:07:00 crc kubenswrapper[4990]: I1203 13:07:00.208666 4990 generic.go:334] "Generic (PLEG): container finished" podID="3c11a50b-b2d3-45be-9999-bcd93b698b1c" containerID="1e319ccf1dff713fbf85874fa9eb86be0a1576b306741a798fc6a6158f4f4748" exitCode=0 Dec 03 13:07:00 crc kubenswrapper[4990]: I1203 13:07:00.208746 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" event={"ID":"3c11a50b-b2d3-45be-9999-bcd93b698b1c","Type":"ContainerDied","Data":"1e319ccf1dff713fbf85874fa9eb86be0a1576b306741a798fc6a6158f4f4748"} Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.606398 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.733894 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-repo-setup-combined-ca-bundle\") pod \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.734685 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvqsd\" (UniqueName: \"kubernetes.io/projected/3c11a50b-b2d3-45be-9999-bcd93b698b1c-kube-api-access-xvqsd\") pod \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.734949 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-inventory\") pod \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.734995 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-ssh-key\") pod \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\" (UID: \"3c11a50b-b2d3-45be-9999-bcd93b698b1c\") " Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.739725 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3c11a50b-b2d3-45be-9999-bcd93b698b1c" (UID: "3c11a50b-b2d3-45be-9999-bcd93b698b1c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.743044 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c11a50b-b2d3-45be-9999-bcd93b698b1c-kube-api-access-xvqsd" (OuterVolumeSpecName: "kube-api-access-xvqsd") pod "3c11a50b-b2d3-45be-9999-bcd93b698b1c" (UID: "3c11a50b-b2d3-45be-9999-bcd93b698b1c"). InnerVolumeSpecName "kube-api-access-xvqsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.763742 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-inventory" (OuterVolumeSpecName: "inventory") pod "3c11a50b-b2d3-45be-9999-bcd93b698b1c" (UID: "3c11a50b-b2d3-45be-9999-bcd93b698b1c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.765189 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3c11a50b-b2d3-45be-9999-bcd93b698b1c" (UID: "3c11a50b-b2d3-45be-9999-bcd93b698b1c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.838842 4990 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.838885 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvqsd\" (UniqueName: \"kubernetes.io/projected/3c11a50b-b2d3-45be-9999-bcd93b698b1c-kube-api-access-xvqsd\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.838907 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:01 crc kubenswrapper[4990]: I1203 13:07:01.838921 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c11a50b-b2d3-45be-9999-bcd93b698b1c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.230094 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" event={"ID":"3c11a50b-b2d3-45be-9999-bcd93b698b1c","Type":"ContainerDied","Data":"ce04577d5d0347a4fd9e0dcd7dbb4b8bb97003abc5a004b090a94284cc260b03"} Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.230157 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p972z" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.230181 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce04577d5d0347a4fd9e0dcd7dbb4b8bb97003abc5a004b090a94284cc260b03" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.316597 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7"] Dec 03 13:07:02 crc kubenswrapper[4990]: E1203 13:07:02.317025 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c11a50b-b2d3-45be-9999-bcd93b698b1c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.317043 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c11a50b-b2d3-45be-9999-bcd93b698b1c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.317243 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c11a50b-b2d3-45be-9999-bcd93b698b1c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.318124 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.320596 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.320851 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.321051 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.321173 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.328562 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7"] Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.349937 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ljq8\" (UniqueName: \"kubernetes.io/projected/748e28f7-7642-4591-bdba-29a79d86f5af-kube-api-access-7ljq8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.350108 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.350185 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.451334 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ljq8\" (UniqueName: \"kubernetes.io/projected/748e28f7-7642-4591-bdba-29a79d86f5af-kube-api-access-7ljq8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.451814 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.451865 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.463635 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.465739 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.469351 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ljq8\" (UniqueName: \"kubernetes.io/projected/748e28f7-7642-4591-bdba-29a79d86f5af-kube-api-access-7ljq8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d2cf7\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:02 crc kubenswrapper[4990]: I1203 13:07:02.649778 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:03 crc kubenswrapper[4990]: I1203 13:07:03.231067 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7"] Dec 03 13:07:03 crc kubenswrapper[4990]: I1203 13:07:03.234774 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:07:03 crc kubenswrapper[4990]: I1203 13:07:03.259112 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" event={"ID":"748e28f7-7642-4591-bdba-29a79d86f5af","Type":"ContainerStarted","Data":"5a0969bdeb35613d06fcb2fd14da237db53f4cc2c491a2c61cba18530e18c9f4"} Dec 03 13:07:03 crc kubenswrapper[4990]: I1203 13:07:03.270897 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:07:03 crc kubenswrapper[4990]: E1203 13:07:03.271213 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:07:04 crc kubenswrapper[4990]: I1203 13:07:04.290447 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" event={"ID":"748e28f7-7642-4591-bdba-29a79d86f5af","Type":"ContainerStarted","Data":"d8bf6f2645f00dab01b1e2d1154aa903da16ba5ba10971662d7ef5cf27724f4c"} Dec 03 13:07:04 crc kubenswrapper[4990]: I1203 13:07:04.292677 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" podStartSLOduration=1.707606264 podStartE2EDuration="2.29265723s" podCreationTimestamp="2025-12-03 13:07:02 +0000 UTC" firstStartedPulling="2025-12-03 13:07:03.234529938 +0000 UTC m=+1771.376441167" lastFinishedPulling="2025-12-03 13:07:03.819580894 +0000 UTC m=+1771.961492133" observedRunningTime="2025-12-03 13:07:04.284996055 +0000 UTC m=+1772.426907294" watchObservedRunningTime="2025-12-03 13:07:04.29265723 +0000 UTC m=+1772.434568459" Dec 03 13:07:07 crc kubenswrapper[4990]: I1203 13:07:07.296102 4990 generic.go:334] "Generic (PLEG): container finished" podID="748e28f7-7642-4591-bdba-29a79d86f5af" containerID="d8bf6f2645f00dab01b1e2d1154aa903da16ba5ba10971662d7ef5cf27724f4c" exitCode=0 Dec 03 13:07:07 crc kubenswrapper[4990]: I1203 13:07:07.296161 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" event={"ID":"748e28f7-7642-4591-bdba-29a79d86f5af","Type":"ContainerDied","Data":"d8bf6f2645f00dab01b1e2d1154aa903da16ba5ba10971662d7ef5cf27724f4c"} Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.789150 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.905761 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-inventory\") pod \"748e28f7-7642-4591-bdba-29a79d86f5af\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.905993 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ljq8\" (UniqueName: \"kubernetes.io/projected/748e28f7-7642-4591-bdba-29a79d86f5af-kube-api-access-7ljq8\") pod \"748e28f7-7642-4591-bdba-29a79d86f5af\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.906039 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-ssh-key\") pod \"748e28f7-7642-4591-bdba-29a79d86f5af\" (UID: \"748e28f7-7642-4591-bdba-29a79d86f5af\") " Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.921691 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/748e28f7-7642-4591-bdba-29a79d86f5af-kube-api-access-7ljq8" (OuterVolumeSpecName: "kube-api-access-7ljq8") pod "748e28f7-7642-4591-bdba-29a79d86f5af" (UID: "748e28f7-7642-4591-bdba-29a79d86f5af"). InnerVolumeSpecName "kube-api-access-7ljq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.939300 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-inventory" (OuterVolumeSpecName: "inventory") pod "748e28f7-7642-4591-bdba-29a79d86f5af" (UID: "748e28f7-7642-4591-bdba-29a79d86f5af"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:07:08 crc kubenswrapper[4990]: I1203 13:07:08.948791 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "748e28f7-7642-4591-bdba-29a79d86f5af" (UID: "748e28f7-7642-4591-bdba-29a79d86f5af"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.009963 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.010047 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ljq8\" (UniqueName: \"kubernetes.io/projected/748e28f7-7642-4591-bdba-29a79d86f5af-kube-api-access-7ljq8\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.010070 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748e28f7-7642-4591-bdba-29a79d86f5af-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.341056 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" event={"ID":"748e28f7-7642-4591-bdba-29a79d86f5af","Type":"ContainerDied","Data":"5a0969bdeb35613d06fcb2fd14da237db53f4cc2c491a2c61cba18530e18c9f4"} Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.341109 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a0969bdeb35613d06fcb2fd14da237db53f4cc2c491a2c61cba18530e18c9f4" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.341171 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d2cf7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.492712 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7"] Dec 03 13:07:09 crc kubenswrapper[4990]: E1203 13:07:09.493199 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="748e28f7-7642-4591-bdba-29a79d86f5af" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.493212 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="748e28f7-7642-4591-bdba-29a79d86f5af" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.493372 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="748e28f7-7642-4591-bdba-29a79d86f5af" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.494051 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.496895 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.497238 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.497999 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.498302 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.508981 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7"] Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.671668 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.671748 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjlrg\" (UniqueName: \"kubernetes.io/projected/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-kube-api-access-zjlrg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.671784 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.671827 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.773927 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.774005 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjlrg\" (UniqueName: \"kubernetes.io/projected/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-kube-api-access-zjlrg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.774025 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.774077 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.779271 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.779395 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.780653 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.793780 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjlrg\" (UniqueName: \"kubernetes.io/projected/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-kube-api-access-zjlrg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:09 crc kubenswrapper[4990]: I1203 13:07:09.818587 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:07:10 crc kubenswrapper[4990]: I1203 13:07:10.425976 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7"] Dec 03 13:07:11 crc kubenswrapper[4990]: I1203 13:07:11.366275 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" event={"ID":"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0","Type":"ContainerStarted","Data":"fb621c753de9c0fc72219283a4949bee795adc3e8f3be2846b1f544d4e33ac10"} Dec 03 13:07:12 crc kubenswrapper[4990]: I1203 13:07:12.376024 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" event={"ID":"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0","Type":"ContainerStarted","Data":"cb9459b5699d7b421cb50d041bd988bd6ab4f5147b3bc6b48b93972723a34eec"} Dec 03 13:07:12 crc kubenswrapper[4990]: I1203 13:07:12.395699 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" podStartSLOduration=1.84404543 podStartE2EDuration="3.395682528s" podCreationTimestamp="2025-12-03 13:07:09 +0000 UTC" firstStartedPulling="2025-12-03 13:07:10.433042165 +0000 UTC m=+1778.574953394" lastFinishedPulling="2025-12-03 13:07:11.984679273 +0000 UTC m=+1780.126590492" observedRunningTime="2025-12-03 13:07:12.389724906 +0000 UTC m=+1780.531636135" watchObservedRunningTime="2025-12-03 13:07:12.395682528 +0000 UTC m=+1780.537593757" Dec 03 13:07:15 crc kubenswrapper[4990]: I1203 13:07:15.265220 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:07:15 crc kubenswrapper[4990]: E1203 13:07:15.265946 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:07:23 crc kubenswrapper[4990]: I1203 13:07:23.751814 4990 scope.go:117] "RemoveContainer" containerID="306f6a8a159f5cb93887336a7515cebebbb3ce0e673f75dca89cd4eb92a69c0b" Dec 03 13:07:27 crc kubenswrapper[4990]: I1203 13:07:27.265210 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:07:27 crc kubenswrapper[4990]: E1203 13:07:27.266502 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:07:40 crc kubenswrapper[4990]: I1203 13:07:40.266098 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:07:40 crc kubenswrapper[4990]: E1203 13:07:40.267749 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:07:55 crc kubenswrapper[4990]: I1203 13:07:55.264539 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:07:55 crc kubenswrapper[4990]: E1203 13:07:55.265475 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:08:08 crc kubenswrapper[4990]: I1203 13:08:08.264529 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:08:08 crc kubenswrapper[4990]: E1203 13:08:08.265284 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:08:23 crc kubenswrapper[4990]: I1203 13:08:23.265908 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:08:23 crc kubenswrapper[4990]: E1203 13:08:23.266978 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:08:23 crc kubenswrapper[4990]: I1203 13:08:23.839739 4990 scope.go:117] "RemoveContainer" containerID="cb6274e5fb2c79285399004898ea1e83c59b93160a26ddfe2f391cfd7b839ccc" Dec 03 13:08:23 crc kubenswrapper[4990]: I1203 13:08:23.866556 4990 scope.go:117] "RemoveContainer" containerID="311f3a153e8a397f7a5dbb6a13f6df38c9cfab799395632057389212ab43d43a" Dec 03 13:08:26 crc kubenswrapper[4990]: I1203 13:08:26.971481 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jvblz"] Dec 03 13:08:26 crc kubenswrapper[4990]: I1203 13:08:26.974099 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:26 crc kubenswrapper[4990]: I1203 13:08:26.988319 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvblz"] Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.128743 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-utilities\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.128795 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-catalog-content\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.128883 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-846zd\" (UniqueName: \"kubernetes.io/projected/480b17b9-90c7-444b-ae00-a0d645668f7c-kube-api-access-846zd\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.230497 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-utilities\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.230555 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-catalog-content\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.230656 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-846zd\" (UniqueName: \"kubernetes.io/projected/480b17b9-90c7-444b-ae00-a0d645668f7c-kube-api-access-846zd\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.231018 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-catalog-content\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.231018 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-utilities\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.256438 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-846zd\" (UniqueName: \"kubernetes.io/projected/480b17b9-90c7-444b-ae00-a0d645668f7c-kube-api-access-846zd\") pod \"redhat-operators-jvblz\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.331633 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:27 crc kubenswrapper[4990]: I1203 13:08:27.781042 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvblz"] Dec 03 13:08:28 crc kubenswrapper[4990]: I1203 13:08:28.092942 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerStarted","Data":"792235a4405d2a270524120e76f10ece705d7ab9753b782cb6eb5e0cd884ac97"} Dec 03 13:08:29 crc kubenswrapper[4990]: I1203 13:08:29.108091 4990 generic.go:334] "Generic (PLEG): container finished" podID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerID="c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866" exitCode=0 Dec 03 13:08:29 crc kubenswrapper[4990]: I1203 13:08:29.108153 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerDied","Data":"c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866"} Dec 03 13:08:31 crc kubenswrapper[4990]: I1203 13:08:31.142149 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerStarted","Data":"3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850"} Dec 03 13:08:34 crc kubenswrapper[4990]: I1203 13:08:34.180069 4990 generic.go:334] "Generic (PLEG): container finished" podID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerID="3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850" exitCode=0 Dec 03 13:08:34 crc kubenswrapper[4990]: I1203 13:08:34.180167 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerDied","Data":"3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850"} Dec 03 13:08:34 crc kubenswrapper[4990]: I1203 13:08:34.264384 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:08:35 crc kubenswrapper[4990]: I1203 13:08:35.190643 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerStarted","Data":"54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6"} Dec 03 13:08:35 crc kubenswrapper[4990]: I1203 13:08:35.194224 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"395d5e989e255143fcd2c670218aa2a6c151ff18e0beac83bd6c077e35197a05"} Dec 03 13:08:35 crc kubenswrapper[4990]: I1203 13:08:35.212872 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jvblz" podStartSLOduration=3.687527216 podStartE2EDuration="9.212851144s" podCreationTimestamp="2025-12-03 13:08:26 +0000 UTC" firstStartedPulling="2025-12-03 13:08:29.11064002 +0000 UTC m=+1857.252551289" lastFinishedPulling="2025-12-03 13:08:34.635963988 +0000 UTC m=+1862.777875217" observedRunningTime="2025-12-03 13:08:35.207157725 +0000 UTC m=+1863.349068964" watchObservedRunningTime="2025-12-03 13:08:35.212851144 +0000 UTC m=+1863.354762373" Dec 03 13:08:37 crc kubenswrapper[4990]: I1203 13:08:37.332376 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:37 crc kubenswrapper[4990]: I1203 13:08:37.333108 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:38 crc kubenswrapper[4990]: I1203 13:08:38.447172 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jvblz" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="registry-server" probeResult="failure" output=< Dec 03 13:08:38 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 13:08:38 crc kubenswrapper[4990]: > Dec 03 13:08:47 crc kubenswrapper[4990]: I1203 13:08:47.401840 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:47 crc kubenswrapper[4990]: I1203 13:08:47.460320 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:47 crc kubenswrapper[4990]: I1203 13:08:47.643337 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvblz"] Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.327877 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jvblz" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="registry-server" containerID="cri-o://54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6" gracePeriod=2 Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.786211 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.916711 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-846zd\" (UniqueName: \"kubernetes.io/projected/480b17b9-90c7-444b-ae00-a0d645668f7c-kube-api-access-846zd\") pod \"480b17b9-90c7-444b-ae00-a0d645668f7c\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.916850 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-catalog-content\") pod \"480b17b9-90c7-444b-ae00-a0d645668f7c\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.916935 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-utilities\") pod \"480b17b9-90c7-444b-ae00-a0d645668f7c\" (UID: \"480b17b9-90c7-444b-ae00-a0d645668f7c\") " Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.917869 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-utilities" (OuterVolumeSpecName: "utilities") pod "480b17b9-90c7-444b-ae00-a0d645668f7c" (UID: "480b17b9-90c7-444b-ae00-a0d645668f7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.919257 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:08:49 crc kubenswrapper[4990]: I1203 13:08:49.922561 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480b17b9-90c7-444b-ae00-a0d645668f7c-kube-api-access-846zd" (OuterVolumeSpecName: "kube-api-access-846zd") pod "480b17b9-90c7-444b-ae00-a0d645668f7c" (UID: "480b17b9-90c7-444b-ae00-a0d645668f7c"). InnerVolumeSpecName "kube-api-access-846zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.020543 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-846zd\" (UniqueName: \"kubernetes.io/projected/480b17b9-90c7-444b-ae00-a0d645668f7c-kube-api-access-846zd\") on node \"crc\" DevicePath \"\"" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.033997 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "480b17b9-90c7-444b-ae00-a0d645668f7c" (UID: "480b17b9-90c7-444b-ae00-a0d645668f7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.123152 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480b17b9-90c7-444b-ae00-a0d645668f7c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.344030 4990 generic.go:334] "Generic (PLEG): container finished" podID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerID="54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6" exitCode=0 Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.344074 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerDied","Data":"54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6"} Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.344113 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvblz" event={"ID":"480b17b9-90c7-444b-ae00-a0d645668f7c","Type":"ContainerDied","Data":"792235a4405d2a270524120e76f10ece705d7ab9753b782cb6eb5e0cd884ac97"} Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.344133 4990 scope.go:117] "RemoveContainer" containerID="54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.344260 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvblz" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.384356 4990 scope.go:117] "RemoveContainer" containerID="3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.384509 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvblz"] Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.394870 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jvblz"] Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.409783 4990 scope.go:117] "RemoveContainer" containerID="c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.453529 4990 scope.go:117] "RemoveContainer" containerID="54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6" Dec 03 13:08:50 crc kubenswrapper[4990]: E1203 13:08:50.454041 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6\": container with ID starting with 54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6 not found: ID does not exist" containerID="54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.454083 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6"} err="failed to get container status \"54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6\": rpc error: code = NotFound desc = could not find container \"54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6\": container with ID starting with 54e8ce655b0d9125df7397344d989181e140bfe8dda299cba056fb532ed49ac6 not found: ID does not exist" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.454108 4990 scope.go:117] "RemoveContainer" containerID="3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850" Dec 03 13:08:50 crc kubenswrapper[4990]: E1203 13:08:50.454669 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850\": container with ID starting with 3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850 not found: ID does not exist" containerID="3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.454710 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850"} err="failed to get container status \"3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850\": rpc error: code = NotFound desc = could not find container \"3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850\": container with ID starting with 3fc145f00652e1349376cf6261a18accb2a67375884e844b6e9ec96b0befc850 not found: ID does not exist" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.454737 4990 scope.go:117] "RemoveContainer" containerID="c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866" Dec 03 13:08:50 crc kubenswrapper[4990]: E1203 13:08:50.455214 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866\": container with ID starting with c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866 not found: ID does not exist" containerID="c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866" Dec 03 13:08:50 crc kubenswrapper[4990]: I1203 13:08:50.455253 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866"} err="failed to get container status \"c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866\": rpc error: code = NotFound desc = could not find container \"c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866\": container with ID starting with c1cd60f7bc905e4370b3405a92085b4e8902f6276af5173fe4b45a81eb37f866 not found: ID does not exist" Dec 03 13:08:52 crc kubenswrapper[4990]: I1203 13:08:52.285671 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" path="/var/lib/kubelet/pods/480b17b9-90c7-444b-ae00-a0d645668f7c/volumes" Dec 03 13:09:43 crc kubenswrapper[4990]: I1203 13:09:43.048112 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3d39-account-create-update-vwp7z"] Dec 03 13:09:43 crc kubenswrapper[4990]: I1203 13:09:43.060160 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3d39-account-create-update-vwp7z"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.038397 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c9da-account-create-update-7zlqj"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.050682 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-dd77-account-create-update-bl6tk"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.062862 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-mtcnj"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.073271 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jdzvw"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.084824 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-dd77-account-create-update-bl6tk"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.098681 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jdzvw"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.109551 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-6fs42"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.119355 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c9da-account-create-update-7zlqj"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.128192 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-mtcnj"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.137995 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-6fs42"] Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.276307 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="140fdc8c-c0ba-4761-a398-cfed9c1179b4" path="/var/lib/kubelet/pods/140fdc8c-c0ba-4761-a398-cfed9c1179b4/volumes" Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.276960 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cd3baf6-2af2-41e9-a3c4-a31fe32588f0" path="/var/lib/kubelet/pods/2cd3baf6-2af2-41e9-a3c4-a31fe32588f0/volumes" Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.277752 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9" path="/var/lib/kubelet/pods/5cec9fbc-5800-48ca-88bc-ecfbb51a1ac9/volumes" Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.278445 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743c06a8-5f6c-4a31-9a4b-08f15170d3b0" path="/var/lib/kubelet/pods/743c06a8-5f6c-4a31-9a4b-08f15170d3b0/volumes" Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.279698 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a03cec11-67cd-483a-929e-003f234f3c6d" path="/var/lib/kubelet/pods/a03cec11-67cd-483a-929e-003f234f3c6d/volumes" Dec 03 13:09:44 crc kubenswrapper[4990]: I1203 13:09:44.280229 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e00bf24a-fcbf-4b46-b502-3c246b290851" path="/var/lib/kubelet/pods/e00bf24a-fcbf-4b46-b502-3c246b290851/volumes" Dec 03 13:09:50 crc kubenswrapper[4990]: I1203 13:09:50.034949 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rx7m6"] Dec 03 13:09:50 crc kubenswrapper[4990]: I1203 13:09:50.047390 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-29e5-account-create-update-t945b"] Dec 03 13:09:50 crc kubenswrapper[4990]: I1203 13:09:50.058727 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rx7m6"] Dec 03 13:09:50 crc kubenswrapper[4990]: I1203 13:09:50.087258 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-29e5-account-create-update-t945b"] Dec 03 13:09:50 crc kubenswrapper[4990]: I1203 13:09:50.278060 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c5047b8-3306-40b5-bbaf-a30c6286662b" path="/var/lib/kubelet/pods/1c5047b8-3306-40b5-bbaf-a30c6286662b/volumes" Dec 03 13:09:50 crc kubenswrapper[4990]: I1203 13:09:50.279168 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28e77ac6-94cf-4a90-b75f-f26c786cb2c6" path="/var/lib/kubelet/pods/28e77ac6-94cf-4a90-b75f-f26c786cb2c6/volumes" Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.040225 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d8c5-account-create-update-qhz9j"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.053563 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5066-account-create-update-w7bck"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.066906 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-gjl5k"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.078329 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-kkvhj"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.086844 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d8c5-account-create-update-qhz9j"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.095788 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5066-account-create-update-w7bck"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.105120 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-gjl5k"] Dec 03 13:09:51 crc kubenswrapper[4990]: I1203 13:09:51.113576 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-kkvhj"] Dec 03 13:09:52 crc kubenswrapper[4990]: I1203 13:09:52.277868 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20984793-ac7c-4ef4-bbf2-cb6e21d03bc7" path="/var/lib/kubelet/pods/20984793-ac7c-4ef4-bbf2-cb6e21d03bc7/volumes" Dec 03 13:09:52 crc kubenswrapper[4990]: I1203 13:09:52.278635 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e3879fb-c88c-4c57-9620-d620812aa5cf" path="/var/lib/kubelet/pods/2e3879fb-c88c-4c57-9620-d620812aa5cf/volumes" Dec 03 13:09:52 crc kubenswrapper[4990]: I1203 13:09:52.279341 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d0ab64-a5d0-4e19-8b39-d38906823542" path="/var/lib/kubelet/pods/45d0ab64-a5d0-4e19-8b39-d38906823542/volumes" Dec 03 13:09:52 crc kubenswrapper[4990]: I1203 13:09:52.280010 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95aafce7-81a5-45a4-aa39-82431c05b665" path="/var/lib/kubelet/pods/95aafce7-81a5-45a4-aa39-82431c05b665/volumes" Dec 03 13:10:23 crc kubenswrapper[4990]: I1203 13:10:23.984882 4990 scope.go:117] "RemoveContainer" containerID="6cdf946732ae067ce4ed96be4c91799ef9763df56be8b2395aae5e057086b982" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.019552 4990 scope.go:117] "RemoveContainer" containerID="9575d22f08298c6a6ad0a7c692b075460e1f831b6a312b85fa105310096cc9a9" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.094848 4990 scope.go:117] "RemoveContainer" containerID="03aefbc52bf21d9e184d73edc83eaa15b8a3e528bf29d6af92fa0dafbf8b49ca" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.121319 4990 scope.go:117] "RemoveContainer" containerID="0325b42ca6ed2d7abce4a6799e178f4a7dc9f71a3fd1c5e630a8af28758f6f80" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.159224 4990 scope.go:117] "RemoveContainer" containerID="7648fd6cf60c1497abcd8aeafd954a237266279d769c2363692d2001a1b4ab39" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.205328 4990 scope.go:117] "RemoveContainer" containerID="1cb1b6ac4dba6a6576c3d28e7424f916b234bbcd8e4ec0c5bedbab7ddc617e8e" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.244738 4990 scope.go:117] "RemoveContainer" containerID="6d485fea3634ef62dde1cc4122194d9c199265ff2dc7e0fa5e0275139fa116c3" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.271785 4990 scope.go:117] "RemoveContainer" containerID="f454a96e8441abaca7cd2ff490cb86afc362ed9140888a2d14eb1f0cf55a7790" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.290435 4990 scope.go:117] "RemoveContainer" containerID="80a36ee5d21d345ea1e61546d3b9de599f81caccc7d2905415897e1f22c630ab" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.315547 4990 scope.go:117] "RemoveContainer" containerID="381554b619f1bfc51a2f37272a046996b23f7b1d933e399fc35229629e1f49a8" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.340016 4990 scope.go:117] "RemoveContainer" containerID="24ca0561482fd601a76f310fd5fdaad86b52740bd4c9f6799180d6d2bd09dc0b" Dec 03 13:10:24 crc kubenswrapper[4990]: I1203 13:10:24.373895 4990 scope.go:117] "RemoveContainer" containerID="3464588fc087417713056f848bac5c229c60657762b721327e4d3682fbc59af6" Dec 03 13:10:25 crc kubenswrapper[4990]: I1203 13:10:25.057224 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-pfvtg"] Dec 03 13:10:25 crc kubenswrapper[4990]: I1203 13:10:25.067493 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-pfvtg"] Dec 03 13:10:26 crc kubenswrapper[4990]: I1203 13:10:26.276289 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb56fae9-35bb-4e75-95b3-6ba3185189aa" path="/var/lib/kubelet/pods/fb56fae9-35bb-4e75-95b3-6ba3185189aa/volumes" Dec 03 13:11:03 crc kubenswrapper[4990]: I1203 13:11:03.285978 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:11:03 crc kubenswrapper[4990]: I1203 13:11:03.286539 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:11:24 crc kubenswrapper[4990]: I1203 13:11:24.590172 4990 scope.go:117] "RemoveContainer" containerID="b8b001866ba17d9f295eb7f115c2e6fe780b3421d5838e9b6e5d141dc09eb30b" Dec 03 13:11:24 crc kubenswrapper[4990]: I1203 13:11:24.632177 4990 scope.go:117] "RemoveContainer" containerID="546a1ef912dde06407c6d739cc39028d386381d501c94ce8033a9875c547fe3b" Dec 03 13:11:24 crc kubenswrapper[4990]: I1203 13:11:24.672971 4990 scope.go:117] "RemoveContainer" containerID="d640fdf5618d25d7277da5d497e23562a7c88e997a561356044c85aa0c949b78" Dec 03 13:11:24 crc kubenswrapper[4990]: I1203 13:11:24.703875 4990 scope.go:117] "RemoveContainer" containerID="c02236eb6c3d6123a966261adcb0298733d380099030b20571666800bee2441e" Dec 03 13:11:33 crc kubenswrapper[4990]: I1203 13:11:33.286358 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:11:33 crc kubenswrapper[4990]: I1203 13:11:33.286877 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:11:37 crc kubenswrapper[4990]: I1203 13:11:37.995313 4990 generic.go:334] "Generic (PLEG): container finished" podID="6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" containerID="cb9459b5699d7b421cb50d041bd988bd6ab4f5147b3bc6b48b93972723a34eec" exitCode=0 Dec 03 13:11:37 crc kubenswrapper[4990]: I1203 13:11:37.995362 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" event={"ID":"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0","Type":"ContainerDied","Data":"cb9459b5699d7b421cb50d041bd988bd6ab4f5147b3bc6b48b93972723a34eec"} Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.410752 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.458026 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-bootstrap-combined-ca-bundle\") pod \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.458127 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-inventory\") pod \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.458309 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjlrg\" (UniqueName: \"kubernetes.io/projected/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-kube-api-access-zjlrg\") pod \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.458335 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-ssh-key\") pod \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\" (UID: \"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0\") " Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.462666 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" (UID: "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.463673 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-kube-api-access-zjlrg" (OuterVolumeSpecName: "kube-api-access-zjlrg") pod "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" (UID: "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0"). InnerVolumeSpecName "kube-api-access-zjlrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.484741 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" (UID: "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.487291 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-inventory" (OuterVolumeSpecName: "inventory") pod "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" (UID: "6c0707b9-5be8-4bf4-870b-ace30c9a9fe0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.561186 4990 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.561232 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.561241 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjlrg\" (UniqueName: \"kubernetes.io/projected/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-kube-api-access-zjlrg\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:39 crc kubenswrapper[4990]: I1203 13:11:39.561249 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c0707b9-5be8-4bf4-870b-ace30c9a9fe0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.029911 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" event={"ID":"6c0707b9-5be8-4bf4-870b-ace30c9a9fe0","Type":"ContainerDied","Data":"fb621c753de9c0fc72219283a4949bee795adc3e8f3be2846b1f544d4e33ac10"} Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.029981 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb621c753de9c0fc72219283a4949bee795adc3e8f3be2846b1f544d4e33ac10" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.030100 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.125035 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr"] Dec 03 13:11:40 crc kubenswrapper[4990]: E1203 13:11:40.125789 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="extract-content" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.125884 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="extract-content" Dec 03 13:11:40 crc kubenswrapper[4990]: E1203 13:11:40.125961 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.126042 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 13:11:40 crc kubenswrapper[4990]: E1203 13:11:40.126160 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="extract-utilities" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.126239 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="extract-utilities" Dec 03 13:11:40 crc kubenswrapper[4990]: E1203 13:11:40.126332 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="registry-server" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.126400 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="registry-server" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.126713 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="480b17b9-90c7-444b-ae00-a0d645668f7c" containerName="registry-server" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.126831 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c0707b9-5be8-4bf4-870b-ace30c9a9fe0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.127815 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.130929 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.131426 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.131608 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.134930 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.136281 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr"] Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.174719 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-889rd\" (UniqueName: \"kubernetes.io/projected/48109f08-9566-4bb0-bc9c-68413ddd32ae-kube-api-access-889rd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.174810 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.174864 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.276665 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.276721 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.277334 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-889rd\" (UniqueName: \"kubernetes.io/projected/48109f08-9566-4bb0-bc9c-68413ddd32ae-kube-api-access-889rd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.281096 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.285988 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.299360 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-889rd\" (UniqueName: \"kubernetes.io/projected/48109f08-9566-4bb0-bc9c-68413ddd32ae-kube-api-access-889rd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-57tqr\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.460028 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:11:40 crc kubenswrapper[4990]: I1203 13:11:40.974707 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr"] Dec 03 13:11:41 crc kubenswrapper[4990]: I1203 13:11:41.043082 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" event={"ID":"48109f08-9566-4bb0-bc9c-68413ddd32ae","Type":"ContainerStarted","Data":"d7a12c762b07cdc7362164508ed4b5bde257ad77df210e8943306d3d33aa5bdf"} Dec 03 13:11:42 crc kubenswrapper[4990]: I1203 13:11:42.063084 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" event={"ID":"48109f08-9566-4bb0-bc9c-68413ddd32ae","Type":"ContainerStarted","Data":"b11aa2e7e3d431ead6df0854d36b961d56cf7c6352affc9756ca27435af28f35"} Dec 03 13:11:42 crc kubenswrapper[4990]: I1203 13:11:42.080381 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" podStartSLOduration=1.616730065 podStartE2EDuration="2.080365811s" podCreationTimestamp="2025-12-03 13:11:40 +0000 UTC" firstStartedPulling="2025-12-03 13:11:40.978118824 +0000 UTC m=+2049.120030053" lastFinishedPulling="2025-12-03 13:11:41.44175457 +0000 UTC m=+2049.583665799" observedRunningTime="2025-12-03 13:11:42.07727498 +0000 UTC m=+2050.219186229" watchObservedRunningTime="2025-12-03 13:11:42.080365811 +0000 UTC m=+2050.222277040" Dec 03 13:11:48 crc kubenswrapper[4990]: I1203 13:11:48.047295 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-p924d"] Dec 03 13:11:48 crc kubenswrapper[4990]: I1203 13:11:48.063513 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-p924d"] Dec 03 13:11:48 crc kubenswrapper[4990]: I1203 13:11:48.281250 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="128f332d-3a7a-4e69-a5ec-5b4bca0d7c59" path="/var/lib/kubelet/pods/128f332d-3a7a-4e69-a5ec-5b4bca0d7c59/volumes" Dec 03 13:11:53 crc kubenswrapper[4990]: I1203 13:11:53.035790 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-9n7xj"] Dec 03 13:11:53 crc kubenswrapper[4990]: I1203 13:11:53.048686 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-9n7xj"] Dec 03 13:11:54 crc kubenswrapper[4990]: I1203 13:11:54.275505 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a" path="/var/lib/kubelet/pods/2e7e0f06-b14f-4fe5-8b02-fa3806f1e52a/volumes" Dec 03 13:12:00 crc kubenswrapper[4990]: I1203 13:12:00.037713 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2kjtg"] Dec 03 13:12:00 crc kubenswrapper[4990]: I1203 13:12:00.045230 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2kjtg"] Dec 03 13:12:00 crc kubenswrapper[4990]: I1203 13:12:00.284003 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0890cf-e6bc-43e0-98de-f37e3aba3d60" path="/var/lib/kubelet/pods/5b0890cf-e6bc-43e0-98de-f37e3aba3d60/volumes" Dec 03 13:12:01 crc kubenswrapper[4990]: I1203 13:12:01.029190 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-psk4l"] Dec 03 13:12:01 crc kubenswrapper[4990]: I1203 13:12:01.040933 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-cqrdj"] Dec 03 13:12:01 crc kubenswrapper[4990]: I1203 13:12:01.050383 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-psk4l"] Dec 03 13:12:01 crc kubenswrapper[4990]: I1203 13:12:01.057948 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-cqrdj"] Dec 03 13:12:02 crc kubenswrapper[4990]: I1203 13:12:02.280740 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f8139e2-22c8-47ce-a03d-db1d0e9e2a29" path="/var/lib/kubelet/pods/2f8139e2-22c8-47ce-a03d-db1d0e9e2a29/volumes" Dec 03 13:12:02 crc kubenswrapper[4990]: I1203 13:12:02.282853 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f50f60-4235-4365-84dd-3cfcf9142ae4" path="/var/lib/kubelet/pods/81f50f60-4235-4365-84dd-3cfcf9142ae4/volumes" Dec 03 13:12:03 crc kubenswrapper[4990]: I1203 13:12:03.286511 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:12:03 crc kubenswrapper[4990]: I1203 13:12:03.286567 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:12:03 crc kubenswrapper[4990]: I1203 13:12:03.286634 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:12:03 crc kubenswrapper[4990]: I1203 13:12:03.287486 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"395d5e989e255143fcd2c670218aa2a6c151ff18e0beac83bd6c077e35197a05"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:12:03 crc kubenswrapper[4990]: I1203 13:12:03.287568 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://395d5e989e255143fcd2c670218aa2a6c151ff18e0beac83bd6c077e35197a05" gracePeriod=600 Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.042248 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-flkbm"] Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.052814 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-flkbm"] Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.277552 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f5bbad8-f863-461f-a580-3989b7d2ceaa" path="/var/lib/kubelet/pods/8f5bbad8-f863-461f-a580-3989b7d2ceaa/volumes" Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.300672 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="395d5e989e255143fcd2c670218aa2a6c151ff18e0beac83bd6c077e35197a05" exitCode=0 Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.300731 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"395d5e989e255143fcd2c670218aa2a6c151ff18e0beac83bd6c077e35197a05"} Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.300780 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b"} Dec 03 13:12:04 crc kubenswrapper[4990]: I1203 13:12:04.300810 4990 scope.go:117] "RemoveContainer" containerID="af2bb4a2315745531c0399ae07f6add9d53280ec43eed17de50f0b4b56e85a9b" Dec 03 13:12:24 crc kubenswrapper[4990]: I1203 13:12:24.762736 4990 scope.go:117] "RemoveContainer" containerID="d49e602e9395e82f917f91b5d9ab1872c584be6da93decd14e0683b3ec32ec7f" Dec 03 13:12:24 crc kubenswrapper[4990]: I1203 13:12:24.788947 4990 scope.go:117] "RemoveContainer" containerID="5dc03a09fcb7e7ece4cd181ace64de8eec00fcd8301f60344f0143c551e50943" Dec 03 13:12:24 crc kubenswrapper[4990]: I1203 13:12:24.838295 4990 scope.go:117] "RemoveContainer" containerID="6ea21dd01851bbc1d1cdd0e941b8c7c9214ebf307d1e384d040cada3a134ce4a" Dec 03 13:12:24 crc kubenswrapper[4990]: I1203 13:12:24.890062 4990 scope.go:117] "RemoveContainer" containerID="aab463dca2b31d447a4e07b95e29921ba560ce57a46aeb6c25e89fd4e399c7ad" Dec 03 13:12:24 crc kubenswrapper[4990]: I1203 13:12:24.933181 4990 scope.go:117] "RemoveContainer" containerID="dac06d20134555dcc54ebef75bc8e29100ebcf015c133b3803b6f4955e61c7df" Dec 03 13:12:24 crc kubenswrapper[4990]: I1203 13:12:24.972552 4990 scope.go:117] "RemoveContainer" containerID="fc33847e50961eb2b69818709bfa216afc7586e78b77349b8c8fc0d7e42959f4" Dec 03 13:12:25 crc kubenswrapper[4990]: I1203 13:12:25.008082 4990 scope.go:117] "RemoveContainer" containerID="6b207d159a2501c336acd5bbdd3a5760fe0d6d0d35132b5cb94f25ed7b5850f7" Dec 03 13:12:25 crc kubenswrapper[4990]: I1203 13:12:25.049669 4990 scope.go:117] "RemoveContainer" containerID="6ff6f94a40fc098f0cc441eb99c4845617257d31c7da670e8018ed042ce12caf" Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.067147 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-54jvk"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.079637 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5857-account-create-update-bf22s"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.091611 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-7mtkf"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.114733 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-54jvk"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.121710 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5857-account-create-update-bf22s"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.128200 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-znfpn"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.134308 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-7mtkf"] Dec 03 13:12:59 crc kubenswrapper[4990]: I1203 13:12:59.140607 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-znfpn"] Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.030434 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6d20-account-create-update-929g9"] Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.037969 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3383-account-create-update-6ggwh"] Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.046417 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3383-account-create-update-6ggwh"] Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.053941 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6d20-account-create-update-929g9"] Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.277147 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b129617-7cab-45cc-9cb0-00fdf619bec2" path="/var/lib/kubelet/pods/1b129617-7cab-45cc-9cb0-00fdf619bec2/volumes" Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.277844 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d0fa5a6-22a0-4550-ba7e-ab7f23efad27" path="/var/lib/kubelet/pods/1d0fa5a6-22a0-4550-ba7e-ab7f23efad27/volumes" Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.278493 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aceea00-aeff-45f4-83ae-c7656a28be74" path="/var/lib/kubelet/pods/2aceea00-aeff-45f4-83ae-c7656a28be74/volumes" Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.279134 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e" path="/var/lib/kubelet/pods/3c4e36ec-4ef9-4773-8ae2-2b0dabf61b3e/volumes" Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.280282 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d4ddf00-b53e-4d9c-bfaa-676226812d86" path="/var/lib/kubelet/pods/8d4ddf00-b53e-4d9c-bfaa-676226812d86/volumes" Dec 03 13:13:00 crc kubenswrapper[4990]: I1203 13:13:00.281086 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a46743-abd7-4cc7-9594-60332bb6aa8d" path="/var/lib/kubelet/pods/a7a46743-abd7-4cc7-9594-60332bb6aa8d/volumes" Dec 03 13:13:25 crc kubenswrapper[4990]: I1203 13:13:25.212409 4990 scope.go:117] "RemoveContainer" containerID="0aaeac0d890447ab06b1675bbc0ecd452feed1f9651d741ea7b8a363829e68d3" Dec 03 13:13:25 crc kubenswrapper[4990]: I1203 13:13:25.240106 4990 scope.go:117] "RemoveContainer" containerID="f5451a9033146b5ec9114fd6279daf4099c6deb267cf5b1ff67747226bdfa42d" Dec 03 13:13:25 crc kubenswrapper[4990]: I1203 13:13:25.309393 4990 scope.go:117] "RemoveContainer" containerID="8a62221fe96a9eb71de6e3730b0bc15c912092e90565f1cbfc86aed4918e7823" Dec 03 13:13:25 crc kubenswrapper[4990]: I1203 13:13:25.350334 4990 scope.go:117] "RemoveContainer" containerID="fee086a7a2fe21abcad12fe048099b83ad057b3b2f7c3b7157205f2850508c83" Dec 03 13:13:25 crc kubenswrapper[4990]: I1203 13:13:25.394344 4990 scope.go:117] "RemoveContainer" containerID="4a51ba2b9557f90c165c3ca6289b1d5342f36d2366e0cad02d656032cf556801" Dec 03 13:13:25 crc kubenswrapper[4990]: I1203 13:13:25.425233 4990 scope.go:117] "RemoveContainer" containerID="b4bdde7dd51e54840effae84404aa7d83e40ded8774a607e6f79223cd7052989" Dec 03 13:13:51 crc kubenswrapper[4990]: I1203 13:13:51.371995 4990 generic.go:334] "Generic (PLEG): container finished" podID="48109f08-9566-4bb0-bc9c-68413ddd32ae" containerID="b11aa2e7e3d431ead6df0854d36b961d56cf7c6352affc9756ca27435af28f35" exitCode=0 Dec 03 13:13:51 crc kubenswrapper[4990]: I1203 13:13:51.372090 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" event={"ID":"48109f08-9566-4bb0-bc9c-68413ddd32ae","Type":"ContainerDied","Data":"b11aa2e7e3d431ead6df0854d36b961d56cf7c6352affc9756ca27435af28f35"} Dec 03 13:13:52 crc kubenswrapper[4990]: I1203 13:13:52.899997 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:13:52 crc kubenswrapper[4990]: I1203 13:13:52.903781 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-inventory\") pod \"48109f08-9566-4bb0-bc9c-68413ddd32ae\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " Dec 03 13:13:52 crc kubenswrapper[4990]: I1203 13:13:52.904877 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-889rd\" (UniqueName: \"kubernetes.io/projected/48109f08-9566-4bb0-bc9c-68413ddd32ae-kube-api-access-889rd\") pod \"48109f08-9566-4bb0-bc9c-68413ddd32ae\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " Dec 03 13:13:52 crc kubenswrapper[4990]: I1203 13:13:52.913090 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48109f08-9566-4bb0-bc9c-68413ddd32ae-kube-api-access-889rd" (OuterVolumeSpecName: "kube-api-access-889rd") pod "48109f08-9566-4bb0-bc9c-68413ddd32ae" (UID: "48109f08-9566-4bb0-bc9c-68413ddd32ae"). InnerVolumeSpecName "kube-api-access-889rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:13:52 crc kubenswrapper[4990]: I1203 13:13:52.954622 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-inventory" (OuterVolumeSpecName: "inventory") pod "48109f08-9566-4bb0-bc9c-68413ddd32ae" (UID: "48109f08-9566-4bb0-bc9c-68413ddd32ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.006963 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-ssh-key\") pod \"48109f08-9566-4bb0-bc9c-68413ddd32ae\" (UID: \"48109f08-9566-4bb0-bc9c-68413ddd32ae\") " Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.007691 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.007720 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-889rd\" (UniqueName: \"kubernetes.io/projected/48109f08-9566-4bb0-bc9c-68413ddd32ae-kube-api-access-889rd\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.036185 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48109f08-9566-4bb0-bc9c-68413ddd32ae" (UID: "48109f08-9566-4bb0-bc9c-68413ddd32ae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.109018 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48109f08-9566-4bb0-bc9c-68413ddd32ae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.395471 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" event={"ID":"48109f08-9566-4bb0-bc9c-68413ddd32ae","Type":"ContainerDied","Data":"d7a12c762b07cdc7362164508ed4b5bde257ad77df210e8943306d3d33aa5bdf"} Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.395511 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7a12c762b07cdc7362164508ed4b5bde257ad77df210e8943306d3d33aa5bdf" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.395559 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-57tqr" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.518908 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb"] Dec 03 13:13:53 crc kubenswrapper[4990]: E1203 13:13:53.519481 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48109f08-9566-4bb0-bc9c-68413ddd32ae" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.519507 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="48109f08-9566-4bb0-bc9c-68413ddd32ae" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.519745 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="48109f08-9566-4bb0-bc9c-68413ddd32ae" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.520522 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.525678 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.525803 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.525875 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.526141 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.544660 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb"] Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.621324 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.621491 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs8qf\" (UniqueName: \"kubernetes.io/projected/7c2c7e25-8434-436c-8e35-61011e98873c-kube-api-access-hs8qf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.621581 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.723737 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.723988 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.724138 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs8qf\" (UniqueName: \"kubernetes.io/projected/7c2c7e25-8434-436c-8e35-61011e98873c-kube-api-access-hs8qf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.729384 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.729646 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.755609 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs8qf\" (UniqueName: \"kubernetes.io/projected/7c2c7e25-8434-436c-8e35-61011e98873c-kube-api-access-hs8qf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wltmb\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:53 crc kubenswrapper[4990]: I1203 13:13:53.841107 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:13:54 crc kubenswrapper[4990]: I1203 13:13:54.385562 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb"] Dec 03 13:13:54 crc kubenswrapper[4990]: I1203 13:13:54.393920 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:13:54 crc kubenswrapper[4990]: I1203 13:13:54.410006 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" event={"ID":"7c2c7e25-8434-436c-8e35-61011e98873c","Type":"ContainerStarted","Data":"32a3d36e73a85d1a658b26a7911aa5f64166f5e04389b455bbe153dc1bca93d9"} Dec 03 13:13:55 crc kubenswrapper[4990]: I1203 13:13:55.422522 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" event={"ID":"7c2c7e25-8434-436c-8e35-61011e98873c","Type":"ContainerStarted","Data":"a8f22cbd9e220b174084bc6a797665e79173f8276b0cec1b46270ef86f1ff039"} Dec 03 13:13:55 crc kubenswrapper[4990]: I1203 13:13:55.444791 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" podStartSLOduration=1.950770436 podStartE2EDuration="2.444767567s" podCreationTimestamp="2025-12-03 13:13:53 +0000 UTC" firstStartedPulling="2025-12-03 13:13:54.393276039 +0000 UTC m=+2182.535187318" lastFinishedPulling="2025-12-03 13:13:54.88727322 +0000 UTC m=+2183.029184449" observedRunningTime="2025-12-03 13:13:55.442904898 +0000 UTC m=+2183.584816157" watchObservedRunningTime="2025-12-03 13:13:55.444767567 +0000 UTC m=+2183.586678836" Dec 03 13:13:57 crc kubenswrapper[4990]: I1203 13:13:57.064535 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vkq7n"] Dec 03 13:13:57 crc kubenswrapper[4990]: I1203 13:13:57.081138 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vkq7n"] Dec 03 13:13:58 crc kubenswrapper[4990]: I1203 13:13:58.275127 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f6005f7-25b3-4e17-bacf-763d1f68e22c" path="/var/lib/kubelet/pods/9f6005f7-25b3-4e17-bacf-763d1f68e22c/volumes" Dec 03 13:14:03 crc kubenswrapper[4990]: I1203 13:14:03.286804 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:14:03 crc kubenswrapper[4990]: I1203 13:14:03.287417 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:14:19 crc kubenswrapper[4990]: I1203 13:14:19.037595 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5szpf"] Dec 03 13:14:19 crc kubenswrapper[4990]: I1203 13:14:19.044965 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5szpf"] Dec 03 13:14:20 crc kubenswrapper[4990]: I1203 13:14:20.282066 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89eeb481-64d0-4b5a-b470-8edd0091c19b" path="/var/lib/kubelet/pods/89eeb481-64d0-4b5a-b470-8edd0091c19b/volumes" Dec 03 13:14:25 crc kubenswrapper[4990]: I1203 13:14:25.576309 4990 scope.go:117] "RemoveContainer" containerID="5a3d5d1b41a54f83b496297d853637e6078bebfdd00162e8f69b807a18d7ef6e" Dec 03 13:14:25 crc kubenswrapper[4990]: I1203 13:14:25.643368 4990 scope.go:117] "RemoveContainer" containerID="133b18a07e1097b63c89fec7ad14818a5236027a276aca7e8f19f8e29508fa23" Dec 03 13:14:27 crc kubenswrapper[4990]: I1203 13:14:27.060558 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h4gl2"] Dec 03 13:14:27 crc kubenswrapper[4990]: I1203 13:14:27.070555 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-h4gl2"] Dec 03 13:14:28 crc kubenswrapper[4990]: I1203 13:14:28.282051 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45e64cfc-87ef-43b1-9b8a-e76314fc514e" path="/var/lib/kubelet/pods/45e64cfc-87ef-43b1-9b8a-e76314fc514e/volumes" Dec 03 13:14:33 crc kubenswrapper[4990]: I1203 13:14:33.286510 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:14:33 crc kubenswrapper[4990]: I1203 13:14:33.287112 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.145391 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4"] Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.147974 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.150962 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.151070 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.169334 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4"] Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.232895 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-config-volume\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.232983 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-secret-volume\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.233011 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqs72\" (UniqueName: \"kubernetes.io/projected/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-kube-api-access-kqs72\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.334127 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-config-volume\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.334209 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-secret-volume\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.334242 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqs72\" (UniqueName: \"kubernetes.io/projected/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-kube-api-access-kqs72\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.335638 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-config-volume\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.340388 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-secret-volume\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.355234 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqs72\" (UniqueName: \"kubernetes.io/projected/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-kube-api-access-kqs72\") pod \"collect-profiles-29412795-8ggb4\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.469591 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:00 crc kubenswrapper[4990]: I1203 13:15:00.912397 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4"] Dec 03 13:15:01 crc kubenswrapper[4990]: I1203 13:15:01.044833 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" event={"ID":"22a53162-4ed1-48f6-bdf2-4dc08f7ba143","Type":"ContainerStarted","Data":"d6a423d303cd5e27dbd1f15adbe614bb7a617b2991b3ab93f3084607dd908c7a"} Dec 03 13:15:02 crc kubenswrapper[4990]: I1203 13:15:02.059768 4990 generic.go:334] "Generic (PLEG): container finished" podID="22a53162-4ed1-48f6-bdf2-4dc08f7ba143" containerID="6baa2e1e5f3d7af054d9629c36642024cff71374a31dbb911267a5d7be325814" exitCode=0 Dec 03 13:15:02 crc kubenswrapper[4990]: I1203 13:15:02.059855 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" event={"ID":"22a53162-4ed1-48f6-bdf2-4dc08f7ba143","Type":"ContainerDied","Data":"6baa2e1e5f3d7af054d9629c36642024cff71374a31dbb911267a5d7be325814"} Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.287047 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.287320 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.287365 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.288069 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.288131 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" gracePeriod=600 Dec 03 13:15:03 crc kubenswrapper[4990]: E1203 13:15:03.409083 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.482089 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.600266 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-config-volume\") pod \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.600383 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-secret-volume\") pod \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.600443 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqs72\" (UniqueName: \"kubernetes.io/projected/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-kube-api-access-kqs72\") pod \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\" (UID: \"22a53162-4ed1-48f6-bdf2-4dc08f7ba143\") " Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.600953 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-config-volume" (OuterVolumeSpecName: "config-volume") pod "22a53162-4ed1-48f6-bdf2-4dc08f7ba143" (UID: "22a53162-4ed1-48f6-bdf2-4dc08f7ba143"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.601221 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.605677 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-kube-api-access-kqs72" (OuterVolumeSpecName: "kube-api-access-kqs72") pod "22a53162-4ed1-48f6-bdf2-4dc08f7ba143" (UID: "22a53162-4ed1-48f6-bdf2-4dc08f7ba143"). InnerVolumeSpecName "kube-api-access-kqs72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.605983 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "22a53162-4ed1-48f6-bdf2-4dc08f7ba143" (UID: "22a53162-4ed1-48f6-bdf2-4dc08f7ba143"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.703133 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:03 crc kubenswrapper[4990]: I1203 13:15:03.703420 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqs72\" (UniqueName: \"kubernetes.io/projected/22a53162-4ed1-48f6-bdf2-4dc08f7ba143-kube-api-access-kqs72\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.096009 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" event={"ID":"22a53162-4ed1-48f6-bdf2-4dc08f7ba143","Type":"ContainerDied","Data":"d6a423d303cd5e27dbd1f15adbe614bb7a617b2991b3ab93f3084607dd908c7a"} Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.096062 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6a423d303cd5e27dbd1f15adbe614bb7a617b2991b3ab93f3084607dd908c7a" Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.096035 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4" Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.100911 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" exitCode=0 Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.100964 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b"} Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.101010 4990 scope.go:117] "RemoveContainer" containerID="395d5e989e255143fcd2c670218aa2a6c151ff18e0beac83bd6c077e35197a05" Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.101904 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:15:04 crc kubenswrapper[4990]: E1203 13:15:04.105597 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.563374 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l"] Dec 03 13:15:04 crc kubenswrapper[4990]: I1203 13:15:04.570430 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412750-l877l"] Dec 03 13:15:05 crc kubenswrapper[4990]: I1203 13:15:05.034028 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-zgltr"] Dec 03 13:15:05 crc kubenswrapper[4990]: I1203 13:15:05.042839 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-zgltr"] Dec 03 13:15:06 crc kubenswrapper[4990]: I1203 13:15:06.280852 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826bd110-77e7-484c-bd48-d95fa1b4a0cd" path="/var/lib/kubelet/pods/826bd110-77e7-484c-bd48-d95fa1b4a0cd/volumes" Dec 03 13:15:06 crc kubenswrapper[4990]: I1203 13:15:06.284330 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae0b69fc-aee2-444e-888d-c18a35a386b5" path="/var/lib/kubelet/pods/ae0b69fc-aee2-444e-888d-c18a35a386b5/volumes" Dec 03 13:15:17 crc kubenswrapper[4990]: I1203 13:15:17.264568 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:15:17 crc kubenswrapper[4990]: E1203 13:15:17.265242 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:15:23 crc kubenswrapper[4990]: I1203 13:15:23.306017 4990 generic.go:334] "Generic (PLEG): container finished" podID="7c2c7e25-8434-436c-8e35-61011e98873c" containerID="a8f22cbd9e220b174084bc6a797665e79173f8276b0cec1b46270ef86f1ff039" exitCode=0 Dec 03 13:15:23 crc kubenswrapper[4990]: I1203 13:15:23.306102 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" event={"ID":"7c2c7e25-8434-436c-8e35-61011e98873c","Type":"ContainerDied","Data":"a8f22cbd9e220b174084bc6a797665e79173f8276b0cec1b46270ef86f1ff039"} Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.816094 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.933950 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-inventory\") pod \"7c2c7e25-8434-436c-8e35-61011e98873c\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.934139 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs8qf\" (UniqueName: \"kubernetes.io/projected/7c2c7e25-8434-436c-8e35-61011e98873c-kube-api-access-hs8qf\") pod \"7c2c7e25-8434-436c-8e35-61011e98873c\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.934283 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-ssh-key\") pod \"7c2c7e25-8434-436c-8e35-61011e98873c\" (UID: \"7c2c7e25-8434-436c-8e35-61011e98873c\") " Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.940907 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c2c7e25-8434-436c-8e35-61011e98873c-kube-api-access-hs8qf" (OuterVolumeSpecName: "kube-api-access-hs8qf") pod "7c2c7e25-8434-436c-8e35-61011e98873c" (UID: "7c2c7e25-8434-436c-8e35-61011e98873c"). InnerVolumeSpecName "kube-api-access-hs8qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.983526 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7c2c7e25-8434-436c-8e35-61011e98873c" (UID: "7c2c7e25-8434-436c-8e35-61011e98873c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:24 crc kubenswrapper[4990]: I1203 13:15:24.984728 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-inventory" (OuterVolumeSpecName: "inventory") pod "7c2c7e25-8434-436c-8e35-61011e98873c" (UID: "7c2c7e25-8434-436c-8e35-61011e98873c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.038403 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.038489 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c2c7e25-8434-436c-8e35-61011e98873c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.038533 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs8qf\" (UniqueName: \"kubernetes.io/projected/7c2c7e25-8434-436c-8e35-61011e98873c-kube-api-access-hs8qf\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.326089 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" event={"ID":"7c2c7e25-8434-436c-8e35-61011e98873c","Type":"ContainerDied","Data":"32a3d36e73a85d1a658b26a7911aa5f64166f5e04389b455bbe153dc1bca93d9"} Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.326136 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32a3d36e73a85d1a658b26a7911aa5f64166f5e04389b455bbe153dc1bca93d9" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.326199 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wltmb" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.418031 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv"] Dec 03 13:15:25 crc kubenswrapper[4990]: E1203 13:15:25.418433 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c2c7e25-8434-436c-8e35-61011e98873c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.418470 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c2c7e25-8434-436c-8e35-61011e98873c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:15:25 crc kubenswrapper[4990]: E1203 13:15:25.418519 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a53162-4ed1-48f6-bdf2-4dc08f7ba143" containerName="collect-profiles" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.418528 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a53162-4ed1-48f6-bdf2-4dc08f7ba143" containerName="collect-profiles" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.418728 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a53162-4ed1-48f6-bdf2-4dc08f7ba143" containerName="collect-profiles" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.418772 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c2c7e25-8434-436c-8e35-61011e98873c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.419543 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.421864 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.422152 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.422627 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.423657 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.437113 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv"] Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.555751 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.555811 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.555873 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkl8d\" (UniqueName: \"kubernetes.io/projected/5207255c-c5b2-470d-aa44-57da1579fe01-kube-api-access-kkl8d\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.658126 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.658175 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.658227 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkl8d\" (UniqueName: \"kubernetes.io/projected/5207255c-c5b2-470d-aa44-57da1579fe01-kube-api-access-kkl8d\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.661875 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.677754 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.683299 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkl8d\" (UniqueName: \"kubernetes.io/projected/5207255c-c5b2-470d-aa44-57da1579fe01-kube-api-access-kkl8d\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.720816 4990 scope.go:117] "RemoveContainer" containerID="9b9832b269b2c2b671caf9164566a394f62c930840b8d2fde81058e9e2ed406c" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.761160 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.793027 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k2ggs"] Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.797165 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.823874 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2ggs"] Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.834156 4990 scope.go:117] "RemoveContainer" containerID="67289d189fa31ade806242db89d99050df375597f4f4c62fe97d4235abbed52b" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.871151 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-catalog-content\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.871314 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-utilities\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.871400 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pndf6\" (UniqueName: \"kubernetes.io/projected/ca041e32-f9d7-4106-8a0f-b17ac86268f4-kube-api-access-pndf6\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.956157 4990 scope.go:117] "RemoveContainer" containerID="6904368d80805855204b8af907c9324de9ad1994c0475ee23365de506dcab7c1" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.976437 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-utilities\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.976780 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pndf6\" (UniqueName: \"kubernetes.io/projected/ca041e32-f9d7-4106-8a0f-b17ac86268f4-kube-api-access-pndf6\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.976967 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-catalog-content\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.977328 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-utilities\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:25 crc kubenswrapper[4990]: I1203 13:15:25.993775 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-catalog-content\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:26 crc kubenswrapper[4990]: I1203 13:15:26.000263 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pndf6\" (UniqueName: \"kubernetes.io/projected/ca041e32-f9d7-4106-8a0f-b17ac86268f4-kube-api-access-pndf6\") pod \"community-operators-k2ggs\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:26 crc kubenswrapper[4990]: I1203 13:15:26.271852 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:26 crc kubenswrapper[4990]: I1203 13:15:26.412232 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv"] Dec 03 13:15:26 crc kubenswrapper[4990]: I1203 13:15:26.783545 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2ggs"] Dec 03 13:15:26 crc kubenswrapper[4990]: W1203 13:15:26.788266 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca041e32_f9d7_4106_8a0f_b17ac86268f4.slice/crio-efc8706d3caf05379548c870be34b503b9cf382410510bb4fa17d6290e2b3b96 WatchSource:0}: Error finding container efc8706d3caf05379548c870be34b503b9cf382410510bb4fa17d6290e2b3b96: Status 404 returned error can't find the container with id efc8706d3caf05379548c870be34b503b9cf382410510bb4fa17d6290e2b3b96 Dec 03 13:15:27 crc kubenswrapper[4990]: I1203 13:15:27.356743 4990 generic.go:334] "Generic (PLEG): container finished" podID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerID="c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e" exitCode=0 Dec 03 13:15:27 crc kubenswrapper[4990]: I1203 13:15:27.356799 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerDied","Data":"c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e"} Dec 03 13:15:27 crc kubenswrapper[4990]: I1203 13:15:27.357087 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerStarted","Data":"efc8706d3caf05379548c870be34b503b9cf382410510bb4fa17d6290e2b3b96"} Dec 03 13:15:27 crc kubenswrapper[4990]: I1203 13:15:27.359136 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" event={"ID":"5207255c-c5b2-470d-aa44-57da1579fe01","Type":"ContainerStarted","Data":"234e6d7a63eee4d04999b1ad088cab6b2285ce2d62ebb126fd8d6e231eac370d"} Dec 03 13:15:27 crc kubenswrapper[4990]: I1203 13:15:27.359189 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" event={"ID":"5207255c-c5b2-470d-aa44-57da1579fe01","Type":"ContainerStarted","Data":"fc140bfc609c6c609fb847d2830fb7db4d902af70a89bced67f88a3811372632"} Dec 03 13:15:27 crc kubenswrapper[4990]: I1203 13:15:27.391284 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" podStartSLOduration=1.926485622 podStartE2EDuration="2.391262295s" podCreationTimestamp="2025-12-03 13:15:25 +0000 UTC" firstStartedPulling="2025-12-03 13:15:26.413828556 +0000 UTC m=+2274.555739775" lastFinishedPulling="2025-12-03 13:15:26.878605199 +0000 UTC m=+2275.020516448" observedRunningTime="2025-12-03 13:15:27.385943194 +0000 UTC m=+2275.527854423" watchObservedRunningTime="2025-12-03 13:15:27.391262295 +0000 UTC m=+2275.533173534" Dec 03 13:15:28 crc kubenswrapper[4990]: I1203 13:15:28.371185 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerStarted","Data":"c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337"} Dec 03 13:15:29 crc kubenswrapper[4990]: I1203 13:15:29.381542 4990 generic.go:334] "Generic (PLEG): container finished" podID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerID="c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337" exitCode=0 Dec 03 13:15:29 crc kubenswrapper[4990]: I1203 13:15:29.381581 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerDied","Data":"c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337"} Dec 03 13:15:30 crc kubenswrapper[4990]: I1203 13:15:30.394694 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerStarted","Data":"c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55"} Dec 03 13:15:30 crc kubenswrapper[4990]: I1203 13:15:30.417567 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k2ggs" podStartSLOduration=2.998862243 podStartE2EDuration="5.417547468s" podCreationTimestamp="2025-12-03 13:15:25 +0000 UTC" firstStartedPulling="2025-12-03 13:15:27.358992044 +0000 UTC m=+2275.500903313" lastFinishedPulling="2025-12-03 13:15:29.777677309 +0000 UTC m=+2277.919588538" observedRunningTime="2025-12-03 13:15:30.412776353 +0000 UTC m=+2278.554687592" watchObservedRunningTime="2025-12-03 13:15:30.417547468 +0000 UTC m=+2278.559458697" Dec 03 13:15:32 crc kubenswrapper[4990]: I1203 13:15:32.272708 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:15:32 crc kubenswrapper[4990]: E1203 13:15:32.273238 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:15:32 crc kubenswrapper[4990]: I1203 13:15:32.411774 4990 generic.go:334] "Generic (PLEG): container finished" podID="5207255c-c5b2-470d-aa44-57da1579fe01" containerID="234e6d7a63eee4d04999b1ad088cab6b2285ce2d62ebb126fd8d6e231eac370d" exitCode=0 Dec 03 13:15:32 crc kubenswrapper[4990]: I1203 13:15:32.411823 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" event={"ID":"5207255c-c5b2-470d-aa44-57da1579fe01","Type":"ContainerDied","Data":"234e6d7a63eee4d04999b1ad088cab6b2285ce2d62ebb126fd8d6e231eac370d"} Dec 03 13:15:33 crc kubenswrapper[4990]: I1203 13:15:33.950258 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.141579 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-ssh-key\") pod \"5207255c-c5b2-470d-aa44-57da1579fe01\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.142019 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-inventory\") pod \"5207255c-c5b2-470d-aa44-57da1579fe01\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.142080 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkl8d\" (UniqueName: \"kubernetes.io/projected/5207255c-c5b2-470d-aa44-57da1579fe01-kube-api-access-kkl8d\") pod \"5207255c-c5b2-470d-aa44-57da1579fe01\" (UID: \"5207255c-c5b2-470d-aa44-57da1579fe01\") " Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.152038 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5207255c-c5b2-470d-aa44-57da1579fe01-kube-api-access-kkl8d" (OuterVolumeSpecName: "kube-api-access-kkl8d") pod "5207255c-c5b2-470d-aa44-57da1579fe01" (UID: "5207255c-c5b2-470d-aa44-57da1579fe01"). InnerVolumeSpecName "kube-api-access-kkl8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.173030 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-inventory" (OuterVolumeSpecName: "inventory") pod "5207255c-c5b2-470d-aa44-57da1579fe01" (UID: "5207255c-c5b2-470d-aa44-57da1579fe01"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.181652 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5207255c-c5b2-470d-aa44-57da1579fe01" (UID: "5207255c-c5b2-470d-aa44-57da1579fe01"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.243910 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.243950 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5207255c-c5b2-470d-aa44-57da1579fe01-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.243963 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkl8d\" (UniqueName: \"kubernetes.io/projected/5207255c-c5b2-470d-aa44-57da1579fe01-kube-api-access-kkl8d\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.435148 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.435263 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv" event={"ID":"5207255c-c5b2-470d-aa44-57da1579fe01","Type":"ContainerDied","Data":"fc140bfc609c6c609fb847d2830fb7db4d902af70a89bced67f88a3811372632"} Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.435308 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc140bfc609c6c609fb847d2830fb7db4d902af70a89bced67f88a3811372632" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.627613 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl"] Dec 03 13:15:34 crc kubenswrapper[4990]: E1203 13:15:34.628256 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5207255c-c5b2-470d-aa44-57da1579fe01" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.628388 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5207255c-c5b2-470d-aa44-57da1579fe01" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.628826 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5207255c-c5b2-470d-aa44-57da1579fe01" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.629673 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.631913 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.632273 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.635125 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.639742 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.642157 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl"] Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.757171 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.757519 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.757710 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rlbs\" (UniqueName: \"kubernetes.io/projected/a777faab-eccb-4a4f-b181-836edc76d745-kube-api-access-7rlbs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.859390 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.859977 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.860225 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rlbs\" (UniqueName: \"kubernetes.io/projected/a777faab-eccb-4a4f-b181-836edc76d745-kube-api-access-7rlbs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.870395 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.870975 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.878629 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rlbs\" (UniqueName: \"kubernetes.io/projected/a777faab-eccb-4a4f-b181-836edc76d745-kube-api-access-7rlbs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-blhdl\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:34 crc kubenswrapper[4990]: I1203 13:15:34.947107 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:15:35 crc kubenswrapper[4990]: I1203 13:15:35.581632 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl"] Dec 03 13:15:36 crc kubenswrapper[4990]: I1203 13:15:36.279715 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:36 crc kubenswrapper[4990]: I1203 13:15:36.280044 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:36 crc kubenswrapper[4990]: I1203 13:15:36.343583 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:36 crc kubenswrapper[4990]: I1203 13:15:36.457832 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" event={"ID":"a777faab-eccb-4a4f-b181-836edc76d745","Type":"ContainerStarted","Data":"3d1c44ce063494e9a598e70645451335c4660906d7f7564db2ff0dee2a0f9350"} Dec 03 13:15:36 crc kubenswrapper[4990]: I1203 13:15:36.513045 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:36 crc kubenswrapper[4990]: I1203 13:15:36.583832 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2ggs"] Dec 03 13:15:37 crc kubenswrapper[4990]: I1203 13:15:37.495498 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" event={"ID":"a777faab-eccb-4a4f-b181-836edc76d745","Type":"ContainerStarted","Data":"4e074bb03c2fc78b8dd7dfa7a6df834eefe6f931e769b5b97c9e639e72b19c91"} Dec 03 13:15:37 crc kubenswrapper[4990]: I1203 13:15:37.518843 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" podStartSLOduration=2.611937685 podStartE2EDuration="3.518825363s" podCreationTimestamp="2025-12-03 13:15:34 +0000 UTC" firstStartedPulling="2025-12-03 13:15:35.581695774 +0000 UTC m=+2283.723607013" lastFinishedPulling="2025-12-03 13:15:36.488583462 +0000 UTC m=+2284.630494691" observedRunningTime="2025-12-03 13:15:37.511367787 +0000 UTC m=+2285.653279016" watchObservedRunningTime="2025-12-03 13:15:37.518825363 +0000 UTC m=+2285.660736592" Dec 03 13:15:38 crc kubenswrapper[4990]: I1203 13:15:38.505326 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k2ggs" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="registry-server" containerID="cri-o://c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55" gracePeriod=2 Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.475470 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.517775 4990 generic.go:334] "Generic (PLEG): container finished" podID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerID="c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55" exitCode=0 Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.517825 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerDied","Data":"c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55"} Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.517854 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2ggs" event={"ID":"ca041e32-f9d7-4106-8a0f-b17ac86268f4","Type":"ContainerDied","Data":"efc8706d3caf05379548c870be34b503b9cf382410510bb4fa17d6290e2b3b96"} Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.517880 4990 scope.go:117] "RemoveContainer" containerID="c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.518032 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2ggs" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.545314 4990 scope.go:117] "RemoveContainer" containerID="c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.565715 4990 scope.go:117] "RemoveContainer" containerID="c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.620347 4990 scope.go:117] "RemoveContainer" containerID="c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55" Dec 03 13:15:39 crc kubenswrapper[4990]: E1203 13:15:39.620863 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55\": container with ID starting with c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55 not found: ID does not exist" containerID="c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.620903 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55"} err="failed to get container status \"c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55\": rpc error: code = NotFound desc = could not find container \"c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55\": container with ID starting with c072ca14413636c63bd887aefecd2265a4028db5c62aa4616f72194e9fa11e55 not found: ID does not exist" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.620932 4990 scope.go:117] "RemoveContainer" containerID="c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337" Dec 03 13:15:39 crc kubenswrapper[4990]: E1203 13:15:39.621297 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337\": container with ID starting with c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337 not found: ID does not exist" containerID="c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.621353 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337"} err="failed to get container status \"c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337\": rpc error: code = NotFound desc = could not find container \"c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337\": container with ID starting with c82ed0c92a5785caac7b9b25ce875455606df9b737d2c81746a2c8605c24a337 not found: ID does not exist" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.621385 4990 scope.go:117] "RemoveContainer" containerID="c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e" Dec 03 13:15:39 crc kubenswrapper[4990]: E1203 13:15:39.621757 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e\": container with ID starting with c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e not found: ID does not exist" containerID="c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.621801 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e"} err="failed to get container status \"c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e\": rpc error: code = NotFound desc = could not find container \"c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e\": container with ID starting with c90816c640acf3e21b7dfeb10875e3a3ede50a32c29ae86e51cd98c625042f7e not found: ID does not exist" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.662101 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-catalog-content\") pod \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.662299 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pndf6\" (UniqueName: \"kubernetes.io/projected/ca041e32-f9d7-4106-8a0f-b17ac86268f4-kube-api-access-pndf6\") pod \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.662336 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-utilities\") pod \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\" (UID: \"ca041e32-f9d7-4106-8a0f-b17ac86268f4\") " Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.663459 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-utilities" (OuterVolumeSpecName: "utilities") pod "ca041e32-f9d7-4106-8a0f-b17ac86268f4" (UID: "ca041e32-f9d7-4106-8a0f-b17ac86268f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.681700 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca041e32-f9d7-4106-8a0f-b17ac86268f4-kube-api-access-pndf6" (OuterVolumeSpecName: "kube-api-access-pndf6") pod "ca041e32-f9d7-4106-8a0f-b17ac86268f4" (UID: "ca041e32-f9d7-4106-8a0f-b17ac86268f4"). InnerVolumeSpecName "kube-api-access-pndf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.748645 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca041e32-f9d7-4106-8a0f-b17ac86268f4" (UID: "ca041e32-f9d7-4106-8a0f-b17ac86268f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.764686 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pndf6\" (UniqueName: \"kubernetes.io/projected/ca041e32-f9d7-4106-8a0f-b17ac86268f4-kube-api-access-pndf6\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.764733 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.764742 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca041e32-f9d7-4106-8a0f-b17ac86268f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.849065 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2ggs"] Dec 03 13:15:39 crc kubenswrapper[4990]: I1203 13:15:39.858927 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k2ggs"] Dec 03 13:15:40 crc kubenswrapper[4990]: I1203 13:15:40.277461 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" path="/var/lib/kubelet/pods/ca041e32-f9d7-4106-8a0f-b17ac86268f4/volumes" Dec 03 13:15:43 crc kubenswrapper[4990]: I1203 13:15:43.264876 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:15:43 crc kubenswrapper[4990]: E1203 13:15:43.265621 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:15:58 crc kubenswrapper[4990]: I1203 13:15:58.265037 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:15:58 crc kubenswrapper[4990]: E1203 13:15:58.268271 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.915172 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fc7lg"] Dec 03 13:16:12 crc kubenswrapper[4990]: E1203 13:16:12.917272 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="extract-content" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.917348 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="extract-content" Dec 03 13:16:12 crc kubenswrapper[4990]: E1203 13:16:12.917420 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="extract-utilities" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.917495 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="extract-utilities" Dec 03 13:16:12 crc kubenswrapper[4990]: E1203 13:16:12.917573 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="registry-server" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.917623 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="registry-server" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.917854 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca041e32-f9d7-4106-8a0f-b17ac86268f4" containerName="registry-server" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.919570 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:12 crc kubenswrapper[4990]: I1203 13:16:12.938473 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fc7lg"] Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.078109 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-catalog-content\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.078190 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-utilities\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.078310 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmnwh\" (UniqueName: \"kubernetes.io/projected/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-kube-api-access-dmnwh\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.179996 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmnwh\" (UniqueName: \"kubernetes.io/projected/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-kube-api-access-dmnwh\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.180123 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-catalog-content\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.180181 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-utilities\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.180868 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-utilities\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.180924 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-catalog-content\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.200805 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmnwh\" (UniqueName: \"kubernetes.io/projected/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-kube-api-access-dmnwh\") pod \"certified-operators-fc7lg\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.247707 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.264614 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:16:13 crc kubenswrapper[4990]: E1203 13:16:13.264843 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.495387 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-87nfv"] Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.497700 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.504744 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87nfv"] Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.689679 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-utilities\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.689802 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-catalog-content\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.689825 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfqns\" (UniqueName: \"kubernetes.io/projected/6e60272d-324e-4597-a21a-b7db6f901b19-kube-api-access-xfqns\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.782088 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fc7lg"] Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.791218 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-catalog-content\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.791257 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqns\" (UniqueName: \"kubernetes.io/projected/6e60272d-324e-4597-a21a-b7db6f901b19-kube-api-access-xfqns\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.791375 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-utilities\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.791758 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-catalog-content\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.791858 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-utilities\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.809672 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfqns\" (UniqueName: \"kubernetes.io/projected/6e60272d-324e-4597-a21a-b7db6f901b19-kube-api-access-xfqns\") pod \"redhat-marketplace-87nfv\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.814854 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:13 crc kubenswrapper[4990]: I1203 13:16:13.894724 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerStarted","Data":"1ec4436e41b6b33c69a4b144ef65033a51362afa783f3f5939cad4b3a04a0567"} Dec 03 13:16:14 crc kubenswrapper[4990]: I1203 13:16:14.295058 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87nfv"] Dec 03 13:16:14 crc kubenswrapper[4990]: I1203 13:16:14.908426 4990 generic.go:334] "Generic (PLEG): container finished" podID="6e60272d-324e-4597-a21a-b7db6f901b19" containerID="cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc" exitCode=0 Dec 03 13:16:14 crc kubenswrapper[4990]: I1203 13:16:14.908687 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87nfv" event={"ID":"6e60272d-324e-4597-a21a-b7db6f901b19","Type":"ContainerDied","Data":"cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc"} Dec 03 13:16:14 crc kubenswrapper[4990]: I1203 13:16:14.908789 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87nfv" event={"ID":"6e60272d-324e-4597-a21a-b7db6f901b19","Type":"ContainerStarted","Data":"3f89db48bbb907cddcdbd3ab4075f986de613d4c8e8f4a7b5e478682350d5d1c"} Dec 03 13:16:14 crc kubenswrapper[4990]: I1203 13:16:14.910859 4990 generic.go:334] "Generic (PLEG): container finished" podID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerID="f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53" exitCode=0 Dec 03 13:16:14 crc kubenswrapper[4990]: I1203 13:16:14.910900 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerDied","Data":"f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53"} Dec 03 13:16:15 crc kubenswrapper[4990]: I1203 13:16:15.920903 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerStarted","Data":"8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e"} Dec 03 13:16:16 crc kubenswrapper[4990]: I1203 13:16:16.938729 4990 generic.go:334] "Generic (PLEG): container finished" podID="6e60272d-324e-4597-a21a-b7db6f901b19" containerID="642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac" exitCode=0 Dec 03 13:16:16 crc kubenswrapper[4990]: I1203 13:16:16.938801 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87nfv" event={"ID":"6e60272d-324e-4597-a21a-b7db6f901b19","Type":"ContainerDied","Data":"642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac"} Dec 03 13:16:16 crc kubenswrapper[4990]: I1203 13:16:16.947988 4990 generic.go:334] "Generic (PLEG): container finished" podID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerID="8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e" exitCode=0 Dec 03 13:16:16 crc kubenswrapper[4990]: I1203 13:16:16.948136 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerDied","Data":"8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e"} Dec 03 13:16:18 crc kubenswrapper[4990]: I1203 13:16:18.985728 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87nfv" event={"ID":"6e60272d-324e-4597-a21a-b7db6f901b19","Type":"ContainerStarted","Data":"211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556"} Dec 03 13:16:18 crc kubenswrapper[4990]: I1203 13:16:18.994555 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerStarted","Data":"a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d"} Dec 03 13:16:19 crc kubenswrapper[4990]: I1203 13:16:19.018375 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-87nfv" podStartSLOduration=2.749575614 podStartE2EDuration="6.01835863s" podCreationTimestamp="2025-12-03 13:16:13 +0000 UTC" firstStartedPulling="2025-12-03 13:16:14.910338778 +0000 UTC m=+2323.052250007" lastFinishedPulling="2025-12-03 13:16:18.179121774 +0000 UTC m=+2326.321033023" observedRunningTime="2025-12-03 13:16:19.006836805 +0000 UTC m=+2327.148748044" watchObservedRunningTime="2025-12-03 13:16:19.01835863 +0000 UTC m=+2327.160269859" Dec 03 13:16:19 crc kubenswrapper[4990]: I1203 13:16:19.042604 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fc7lg" podStartSLOduration=3.839570755 podStartE2EDuration="7.042585498s" podCreationTimestamp="2025-12-03 13:16:12 +0000 UTC" firstStartedPulling="2025-12-03 13:16:14.913412219 +0000 UTC m=+2323.055323488" lastFinishedPulling="2025-12-03 13:16:18.116426962 +0000 UTC m=+2326.258338231" observedRunningTime="2025-12-03 13:16:19.034527786 +0000 UTC m=+2327.176439015" watchObservedRunningTime="2025-12-03 13:16:19.042585498 +0000 UTC m=+2327.184496727" Dec 03 13:16:23 crc kubenswrapper[4990]: I1203 13:16:23.247938 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:23 crc kubenswrapper[4990]: I1203 13:16:23.250500 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:23 crc kubenswrapper[4990]: I1203 13:16:23.293203 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:23 crc kubenswrapper[4990]: I1203 13:16:23.815898 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:23 crc kubenswrapper[4990]: I1203 13:16:23.815972 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:23 crc kubenswrapper[4990]: I1203 13:16:23.909003 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:24 crc kubenswrapper[4990]: I1203 13:16:24.065034 4990 generic.go:334] "Generic (PLEG): container finished" podID="a777faab-eccb-4a4f-b181-836edc76d745" containerID="4e074bb03c2fc78b8dd7dfa7a6df834eefe6f931e769b5b97c9e639e72b19c91" exitCode=0 Dec 03 13:16:24 crc kubenswrapper[4990]: I1203 13:16:24.065120 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" event={"ID":"a777faab-eccb-4a4f-b181-836edc76d745","Type":"ContainerDied","Data":"4e074bb03c2fc78b8dd7dfa7a6df834eefe6f931e769b5b97c9e639e72b19c91"} Dec 03 13:16:24 crc kubenswrapper[4990]: I1203 13:16:24.157697 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:24 crc kubenswrapper[4990]: I1203 13:16:24.159290 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.481654 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87nfv"] Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.590631 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.782158 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rlbs\" (UniqueName: \"kubernetes.io/projected/a777faab-eccb-4a4f-b181-836edc76d745-kube-api-access-7rlbs\") pod \"a777faab-eccb-4a4f-b181-836edc76d745\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.782603 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-inventory\") pod \"a777faab-eccb-4a4f-b181-836edc76d745\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.782799 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-ssh-key\") pod \"a777faab-eccb-4a4f-b181-836edc76d745\" (UID: \"a777faab-eccb-4a4f-b181-836edc76d745\") " Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.790119 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a777faab-eccb-4a4f-b181-836edc76d745-kube-api-access-7rlbs" (OuterVolumeSpecName: "kube-api-access-7rlbs") pod "a777faab-eccb-4a4f-b181-836edc76d745" (UID: "a777faab-eccb-4a4f-b181-836edc76d745"). InnerVolumeSpecName "kube-api-access-7rlbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.810114 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a777faab-eccb-4a4f-b181-836edc76d745" (UID: "a777faab-eccb-4a4f-b181-836edc76d745"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.829854 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-inventory" (OuterVolumeSpecName: "inventory") pod "a777faab-eccb-4a4f-b181-836edc76d745" (UID: "a777faab-eccb-4a4f-b181-836edc76d745"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.884624 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rlbs\" (UniqueName: \"kubernetes.io/projected/a777faab-eccb-4a4f-b181-836edc76d745-kube-api-access-7rlbs\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.884670 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:25 crc kubenswrapper[4990]: I1203 13:16:25.884690 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a777faab-eccb-4a4f-b181-836edc76d745-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.088860 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" event={"ID":"a777faab-eccb-4a4f-b181-836edc76d745","Type":"ContainerDied","Data":"3d1c44ce063494e9a598e70645451335c4660906d7f7564db2ff0dee2a0f9350"} Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.088961 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d1c44ce063494e9a598e70645451335c4660906d7f7564db2ff0dee2a0f9350" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.089114 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-blhdl" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.089859 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-87nfv" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="registry-server" containerID="cri-o://211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556" gracePeriod=2 Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.214003 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm"] Dec 03 13:16:26 crc kubenswrapper[4990]: E1203 13:16:26.214588 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a777faab-eccb-4a4f-b181-836edc76d745" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.214613 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a777faab-eccb-4a4f-b181-836edc76d745" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.214894 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="a777faab-eccb-4a4f-b181-836edc76d745" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.215760 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.219949 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.220206 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.220356 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.220408 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.225958 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm"] Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.300577 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvds4\" (UniqueName: \"kubernetes.io/projected/8c311218-88d9-4022-8962-8ff68d03fa29-kube-api-access-kvds4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.300922 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.300987 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.410840 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvds4\" (UniqueName: \"kubernetes.io/projected/8c311218-88d9-4022-8962-8ff68d03fa29-kube-api-access-kvds4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.410913 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.410944 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.415782 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.417285 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.429740 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvds4\" (UniqueName: \"kubernetes.io/projected/8c311218-88d9-4022-8962-8ff68d03fa29-kube-api-access-kvds4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.480099 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fc7lg"] Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.574562 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.670771 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.818597 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfqns\" (UniqueName: \"kubernetes.io/projected/6e60272d-324e-4597-a21a-b7db6f901b19-kube-api-access-xfqns\") pod \"6e60272d-324e-4597-a21a-b7db6f901b19\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.819018 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-catalog-content\") pod \"6e60272d-324e-4597-a21a-b7db6f901b19\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.819087 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-utilities\") pod \"6e60272d-324e-4597-a21a-b7db6f901b19\" (UID: \"6e60272d-324e-4597-a21a-b7db6f901b19\") " Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.820335 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-utilities" (OuterVolumeSpecName: "utilities") pod "6e60272d-324e-4597-a21a-b7db6f901b19" (UID: "6e60272d-324e-4597-a21a-b7db6f901b19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.824683 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e60272d-324e-4597-a21a-b7db6f901b19-kube-api-access-xfqns" (OuterVolumeSpecName: "kube-api-access-xfqns") pod "6e60272d-324e-4597-a21a-b7db6f901b19" (UID: "6e60272d-324e-4597-a21a-b7db6f901b19"). InnerVolumeSpecName "kube-api-access-xfqns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.839774 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e60272d-324e-4597-a21a-b7db6f901b19" (UID: "6e60272d-324e-4597-a21a-b7db6f901b19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.921427 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfqns\" (UniqueName: \"kubernetes.io/projected/6e60272d-324e-4597-a21a-b7db6f901b19-kube-api-access-xfqns\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.921474 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:26 crc kubenswrapper[4990]: I1203 13:16:26.921484 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e60272d-324e-4597-a21a-b7db6f901b19-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.104217 4990 generic.go:334] "Generic (PLEG): container finished" podID="6e60272d-324e-4597-a21a-b7db6f901b19" containerID="211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556" exitCode=0 Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.104277 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87nfv" event={"ID":"6e60272d-324e-4597-a21a-b7db6f901b19","Type":"ContainerDied","Data":"211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556"} Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.104336 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87nfv" event={"ID":"6e60272d-324e-4597-a21a-b7db6f901b19","Type":"ContainerDied","Data":"3f89db48bbb907cddcdbd3ab4075f986de613d4c8e8f4a7b5e478682350d5d1c"} Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.104360 4990 scope.go:117] "RemoveContainer" containerID="211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.104375 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87nfv" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.104537 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fc7lg" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="registry-server" containerID="cri-o://a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d" gracePeriod=2 Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.153652 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87nfv"] Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.157061 4990 scope.go:117] "RemoveContainer" containerID="642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.161760 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-87nfv"] Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.183694 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm"] Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.209436 4990 scope.go:117] "RemoveContainer" containerID="cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.232298 4990 scope.go:117] "RemoveContainer" containerID="211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556" Dec 03 13:16:27 crc kubenswrapper[4990]: E1203 13:16:27.232729 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556\": container with ID starting with 211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556 not found: ID does not exist" containerID="211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.232761 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556"} err="failed to get container status \"211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556\": rpc error: code = NotFound desc = could not find container \"211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556\": container with ID starting with 211d2c74cf0de837f7d1766f714195999cbb09ba875821d1af133cf16c8ba556 not found: ID does not exist" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.232782 4990 scope.go:117] "RemoveContainer" containerID="642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac" Dec 03 13:16:27 crc kubenswrapper[4990]: E1203 13:16:27.233125 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac\": container with ID starting with 642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac not found: ID does not exist" containerID="642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.233180 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac"} err="failed to get container status \"642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac\": rpc error: code = NotFound desc = could not find container \"642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac\": container with ID starting with 642af14323e68d85e88cceebf9f0fe601373410d03ee7ccd09675f8a53cb8eac not found: ID does not exist" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.233218 4990 scope.go:117] "RemoveContainer" containerID="cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc" Dec 03 13:16:27 crc kubenswrapper[4990]: E1203 13:16:27.233653 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc\": container with ID starting with cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc not found: ID does not exist" containerID="cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.233682 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc"} err="failed to get container status \"cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc\": rpc error: code = NotFound desc = could not find container \"cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc\": container with ID starting with cab9620f04af50df47bd2c480cea23c3ef6653f20a18c109dd90104c0d6210cc not found: ID does not exist" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.264882 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:16:27 crc kubenswrapper[4990]: E1203 13:16:27.265150 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.469802 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.632754 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-catalog-content\") pod \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.632867 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmnwh\" (UniqueName: \"kubernetes.io/projected/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-kube-api-access-dmnwh\") pod \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.633700 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-utilities\") pod \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\" (UID: \"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9\") " Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.634761 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-utilities" (OuterVolumeSpecName: "utilities") pod "371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" (UID: "371e7f0b-d5ed-4298-bac2-2c35f0ca4be9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.637172 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-kube-api-access-dmnwh" (OuterVolumeSpecName: "kube-api-access-dmnwh") pod "371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" (UID: "371e7f0b-d5ed-4298-bac2-2c35f0ca4be9"). InnerVolumeSpecName "kube-api-access-dmnwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.678976 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" (UID: "371e7f0b-d5ed-4298-bac2-2c35f0ca4be9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.736166 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.736248 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:27 crc kubenswrapper[4990]: I1203 13:16:27.736274 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmnwh\" (UniqueName: \"kubernetes.io/projected/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9-kube-api-access-dmnwh\") on node \"crc\" DevicePath \"\"" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.114483 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" event={"ID":"8c311218-88d9-4022-8962-8ff68d03fa29","Type":"ContainerStarted","Data":"eea7144f16cca881d32bed6c7a42a6657fac2175a65c6c9c1b34a3185181be8f"} Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.115004 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" event={"ID":"8c311218-88d9-4022-8962-8ff68d03fa29","Type":"ContainerStarted","Data":"30b0153531249d6cb046013051ea26e4bb674979fe40ee368121f713ffad7df9"} Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.119330 4990 generic.go:334] "Generic (PLEG): container finished" podID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerID="a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d" exitCode=0 Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.119482 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerDied","Data":"a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d"} Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.119629 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fc7lg" event={"ID":"371e7f0b-d5ed-4298-bac2-2c35f0ca4be9","Type":"ContainerDied","Data":"1ec4436e41b6b33c69a4b144ef65033a51362afa783f3f5939cad4b3a04a0567"} Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.119720 4990 scope.go:117] "RemoveContainer" containerID="a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.119907 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fc7lg" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.134945 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" podStartSLOduration=1.61118563 podStartE2EDuration="2.134923724s" podCreationTimestamp="2025-12-03 13:16:26 +0000 UTC" firstStartedPulling="2025-12-03 13:16:27.215582681 +0000 UTC m=+2335.357493900" lastFinishedPulling="2025-12-03 13:16:27.739320775 +0000 UTC m=+2335.881231994" observedRunningTime="2025-12-03 13:16:28.128697397 +0000 UTC m=+2336.270608656" watchObservedRunningTime="2025-12-03 13:16:28.134923724 +0000 UTC m=+2336.276834953" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.159370 4990 scope.go:117] "RemoveContainer" containerID="8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.162049 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fc7lg"] Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.171899 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fc7lg"] Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.189704 4990 scope.go:117] "RemoveContainer" containerID="f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.208981 4990 scope.go:117] "RemoveContainer" containerID="a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d" Dec 03 13:16:28 crc kubenswrapper[4990]: E1203 13:16:28.209605 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d\": container with ID starting with a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d not found: ID does not exist" containerID="a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.209644 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d"} err="failed to get container status \"a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d\": rpc error: code = NotFound desc = could not find container \"a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d\": container with ID starting with a7e4a6336c278c4dc5d050bd4fcac1cc0730d9ff13ba989853ad1b0926dcd02d not found: ID does not exist" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.209675 4990 scope.go:117] "RemoveContainer" containerID="8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e" Dec 03 13:16:28 crc kubenswrapper[4990]: E1203 13:16:28.210059 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e\": container with ID starting with 8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e not found: ID does not exist" containerID="8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.210089 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e"} err="failed to get container status \"8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e\": rpc error: code = NotFound desc = could not find container \"8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e\": container with ID starting with 8eb850a3bdaf79798f976845529ce565d99ed7cef47516daf0200d741bf48f9e not found: ID does not exist" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.210118 4990 scope.go:117] "RemoveContainer" containerID="f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53" Dec 03 13:16:28 crc kubenswrapper[4990]: E1203 13:16:28.211592 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53\": container with ID starting with f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53 not found: ID does not exist" containerID="f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.211632 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53"} err="failed to get container status \"f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53\": rpc error: code = NotFound desc = could not find container \"f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53\": container with ID starting with f7d015fa13a0edeb6f2b868a0fa9db13133235cde19c3368abcfbebb604e8c53 not found: ID does not exist" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.278689 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" path="/var/lib/kubelet/pods/371e7f0b-d5ed-4298-bac2-2c35f0ca4be9/volumes" Dec 03 13:16:28 crc kubenswrapper[4990]: I1203 13:16:28.279647 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" path="/var/lib/kubelet/pods/6e60272d-324e-4597-a21a-b7db6f901b19/volumes" Dec 03 13:16:38 crc kubenswrapper[4990]: I1203 13:16:38.265291 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:16:38 crc kubenswrapper[4990]: E1203 13:16:38.266169 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:16:52 crc kubenswrapper[4990]: I1203 13:16:52.271374 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:16:52 crc kubenswrapper[4990]: E1203 13:16:52.273226 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:17:03 crc kubenswrapper[4990]: I1203 13:17:03.265054 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:17:03 crc kubenswrapper[4990]: E1203 13:17:03.265959 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:17:18 crc kubenswrapper[4990]: I1203 13:17:18.264540 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:17:18 crc kubenswrapper[4990]: E1203 13:17:18.265306 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:17:21 crc kubenswrapper[4990]: I1203 13:17:21.667548 4990 generic.go:334] "Generic (PLEG): container finished" podID="8c311218-88d9-4022-8962-8ff68d03fa29" containerID="eea7144f16cca881d32bed6c7a42a6657fac2175a65c6c9c1b34a3185181be8f" exitCode=0 Dec 03 13:17:21 crc kubenswrapper[4990]: I1203 13:17:21.667604 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" event={"ID":"8c311218-88d9-4022-8962-8ff68d03fa29","Type":"ContainerDied","Data":"eea7144f16cca881d32bed6c7a42a6657fac2175a65c6c9c1b34a3185181be8f"} Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.090658 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.177686 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvds4\" (UniqueName: \"kubernetes.io/projected/8c311218-88d9-4022-8962-8ff68d03fa29-kube-api-access-kvds4\") pod \"8c311218-88d9-4022-8962-8ff68d03fa29\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.177781 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-ssh-key\") pod \"8c311218-88d9-4022-8962-8ff68d03fa29\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.177839 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-inventory\") pod \"8c311218-88d9-4022-8962-8ff68d03fa29\" (UID: \"8c311218-88d9-4022-8962-8ff68d03fa29\") " Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.183345 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c311218-88d9-4022-8962-8ff68d03fa29-kube-api-access-kvds4" (OuterVolumeSpecName: "kube-api-access-kvds4") pod "8c311218-88d9-4022-8962-8ff68d03fa29" (UID: "8c311218-88d9-4022-8962-8ff68d03fa29"). InnerVolumeSpecName "kube-api-access-kvds4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.203068 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8c311218-88d9-4022-8962-8ff68d03fa29" (UID: "8c311218-88d9-4022-8962-8ff68d03fa29"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.204416 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-inventory" (OuterVolumeSpecName: "inventory") pod "8c311218-88d9-4022-8962-8ff68d03fa29" (UID: "8c311218-88d9-4022-8962-8ff68d03fa29"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.280764 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvds4\" (UniqueName: \"kubernetes.io/projected/8c311218-88d9-4022-8962-8ff68d03fa29-kube-api-access-kvds4\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.281111 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.281123 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c311218-88d9-4022-8962-8ff68d03fa29-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.684159 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" event={"ID":"8c311218-88d9-4022-8962-8ff68d03fa29","Type":"ContainerDied","Data":"30b0153531249d6cb046013051ea26e4bb674979fe40ee368121f713ffad7df9"} Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.684203 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30b0153531249d6cb046013051ea26e4bb674979fe40ee368121f713ffad7df9" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.684251 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782224 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ntk2h"] Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782715 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c311218-88d9-4022-8962-8ff68d03fa29" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782739 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c311218-88d9-4022-8962-8ff68d03fa29" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782757 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="registry-server" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782765 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="registry-server" Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782776 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="registry-server" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782785 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="registry-server" Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782814 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="extract-utilities" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782821 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="extract-utilities" Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782834 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="extract-content" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782841 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="extract-content" Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782852 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="extract-utilities" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782859 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="extract-utilities" Dec 03 13:17:23 crc kubenswrapper[4990]: E1203 13:17:23.782874 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="extract-content" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.782881 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="extract-content" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.783278 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c311218-88d9-4022-8962-8ff68d03fa29" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.783310 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="371e7f0b-d5ed-4298-bac2-2c35f0ca4be9" containerName="registry-server" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.783324 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e60272d-324e-4597-a21a-b7db6f901b19" containerName="registry-server" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.784310 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.790210 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.790506 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.790754 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.791021 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.805095 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ntk2h"] Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.893595 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.893667 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.893783 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl8r9\" (UniqueName: \"kubernetes.io/projected/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-kube-api-access-wl8r9\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.996261 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl8r9\" (UniqueName: \"kubernetes.io/projected/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-kube-api-access-wl8r9\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.996495 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:23 crc kubenswrapper[4990]: I1203 13:17:23.996525 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:24 crc kubenswrapper[4990]: I1203 13:17:24.003625 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:24 crc kubenswrapper[4990]: I1203 13:17:24.003738 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:24 crc kubenswrapper[4990]: I1203 13:17:24.013750 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl8r9\" (UniqueName: \"kubernetes.io/projected/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-kube-api-access-wl8r9\") pod \"ssh-known-hosts-edpm-deployment-ntk2h\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:24 crc kubenswrapper[4990]: I1203 13:17:24.117978 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:24 crc kubenswrapper[4990]: I1203 13:17:24.630787 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ntk2h"] Dec 03 13:17:24 crc kubenswrapper[4990]: I1203 13:17:24.694548 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" event={"ID":"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf","Type":"ContainerStarted","Data":"41494c5319d6851821e3719b9cb10716b343166f0301d37668b33d6fa1cae7bc"} Dec 03 13:17:25 crc kubenswrapper[4990]: I1203 13:17:25.705003 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" event={"ID":"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf","Type":"ContainerStarted","Data":"cb84872e2ec5ef922c2a4053e90d1f6483798f6b3b3281b02fc1adb0c77008e9"} Dec 03 13:17:25 crc kubenswrapper[4990]: I1203 13:17:25.727128 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" podStartSLOduration=1.981222329 podStartE2EDuration="2.72710668s" podCreationTimestamp="2025-12-03 13:17:23 +0000 UTC" firstStartedPulling="2025-12-03 13:17:24.644483199 +0000 UTC m=+2392.786394428" lastFinishedPulling="2025-12-03 13:17:25.39036754 +0000 UTC m=+2393.532278779" observedRunningTime="2025-12-03 13:17:25.717046041 +0000 UTC m=+2393.858957320" watchObservedRunningTime="2025-12-03 13:17:25.72710668 +0000 UTC m=+2393.869017919" Dec 03 13:17:32 crc kubenswrapper[4990]: I1203 13:17:32.269777 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:17:32 crc kubenswrapper[4990]: E1203 13:17:32.270672 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:17:32 crc kubenswrapper[4990]: I1203 13:17:32.773220 4990 generic.go:334] "Generic (PLEG): container finished" podID="dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" containerID="cb84872e2ec5ef922c2a4053e90d1f6483798f6b3b3281b02fc1adb0c77008e9" exitCode=0 Dec 03 13:17:32 crc kubenswrapper[4990]: I1203 13:17:32.773309 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" event={"ID":"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf","Type":"ContainerDied","Data":"cb84872e2ec5ef922c2a4053e90d1f6483798f6b3b3281b02fc1adb0c77008e9"} Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.198174 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.285973 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl8r9\" (UniqueName: \"kubernetes.io/projected/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-kube-api-access-wl8r9\") pod \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.286392 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-ssh-key-openstack-edpm-ipam\") pod \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.286430 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-inventory-0\") pod \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\" (UID: \"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf\") " Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.291471 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-kube-api-access-wl8r9" (OuterVolumeSpecName: "kube-api-access-wl8r9") pod "dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" (UID: "dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf"). InnerVolumeSpecName "kube-api-access-wl8r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.317670 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" (UID: "dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.319089 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" (UID: "dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.389273 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.389312 4990 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.389326 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl8r9\" (UniqueName: \"kubernetes.io/projected/dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf-kube-api-access-wl8r9\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.790110 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" event={"ID":"dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf","Type":"ContainerDied","Data":"41494c5319d6851821e3719b9cb10716b343166f0301d37668b33d6fa1cae7bc"} Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.790160 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41494c5319d6851821e3719b9cb10716b343166f0301d37668b33d6fa1cae7bc" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.790190 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ntk2h" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.866271 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv"] Dec 03 13:17:34 crc kubenswrapper[4990]: E1203 13:17:34.866695 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" containerName="ssh-known-hosts-edpm-deployment" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.866712 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" containerName="ssh-known-hosts-edpm-deployment" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.866897 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf" containerName="ssh-known-hosts-edpm-deployment" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.867503 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.869643 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.870226 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.870421 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.875019 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.889644 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv"] Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.903540 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x85g8\" (UniqueName: \"kubernetes.io/projected/58dc37ce-d8bf-46f0-9338-43620480d0dd-kube-api-access-x85g8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.903787 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:34 crc kubenswrapper[4990]: I1203 13:17:34.903998 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.006559 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x85g8\" (UniqueName: \"kubernetes.io/projected/58dc37ce-d8bf-46f0-9338-43620480d0dd-kube-api-access-x85g8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.006615 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.006641 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.012133 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.014869 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.028829 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x85g8\" (UniqueName: \"kubernetes.io/projected/58dc37ce-d8bf-46f0-9338-43620480d0dd-kube-api-access-x85g8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hbnrv\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.185242 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.690619 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv"] Dec 03 13:17:35 crc kubenswrapper[4990]: W1203 13:17:35.699524 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58dc37ce_d8bf_46f0_9338_43620480d0dd.slice/crio-f44291a46cef18acdb2a7b2e5088f0a28891df0afac31ae96557e525e9732b94 WatchSource:0}: Error finding container f44291a46cef18acdb2a7b2e5088f0a28891df0afac31ae96557e525e9732b94: Status 404 returned error can't find the container with id f44291a46cef18acdb2a7b2e5088f0a28891df0afac31ae96557e525e9732b94 Dec 03 13:17:35 crc kubenswrapper[4990]: I1203 13:17:35.799282 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" event={"ID":"58dc37ce-d8bf-46f0-9338-43620480d0dd","Type":"ContainerStarted","Data":"f44291a46cef18acdb2a7b2e5088f0a28891df0afac31ae96557e525e9732b94"} Dec 03 13:17:37 crc kubenswrapper[4990]: I1203 13:17:37.818065 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" event={"ID":"58dc37ce-d8bf-46f0-9338-43620480d0dd","Type":"ContainerStarted","Data":"88b1ec74551d8dd874d11d88b56d58eda9f98132ab99ad531d7d96df4579b11d"} Dec 03 13:17:37 crc kubenswrapper[4990]: I1203 13:17:37.839756 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" podStartSLOduration=2.880747827 podStartE2EDuration="3.839735783s" podCreationTimestamp="2025-12-03 13:17:34 +0000 UTC" firstStartedPulling="2025-12-03 13:17:35.701262968 +0000 UTC m=+2403.843174197" lastFinishedPulling="2025-12-03 13:17:36.660250924 +0000 UTC m=+2404.802162153" observedRunningTime="2025-12-03 13:17:37.837244366 +0000 UTC m=+2405.979155605" watchObservedRunningTime="2025-12-03 13:17:37.839735783 +0000 UTC m=+2405.981647012" Dec 03 13:17:43 crc kubenswrapper[4990]: I1203 13:17:43.264879 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:17:43 crc kubenswrapper[4990]: E1203 13:17:43.265656 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:17:44 crc kubenswrapper[4990]: I1203 13:17:44.905166 4990 generic.go:334] "Generic (PLEG): container finished" podID="58dc37ce-d8bf-46f0-9338-43620480d0dd" containerID="88b1ec74551d8dd874d11d88b56d58eda9f98132ab99ad531d7d96df4579b11d" exitCode=0 Dec 03 13:17:44 crc kubenswrapper[4990]: I1203 13:17:44.905252 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" event={"ID":"58dc37ce-d8bf-46f0-9338-43620480d0dd","Type":"ContainerDied","Data":"88b1ec74551d8dd874d11d88b56d58eda9f98132ab99ad531d7d96df4579b11d"} Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.295833 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.385268 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-inventory\") pod \"58dc37ce-d8bf-46f0-9338-43620480d0dd\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.385767 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x85g8\" (UniqueName: \"kubernetes.io/projected/58dc37ce-d8bf-46f0-9338-43620480d0dd-kube-api-access-x85g8\") pod \"58dc37ce-d8bf-46f0-9338-43620480d0dd\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.385799 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key\") pod \"58dc37ce-d8bf-46f0-9338-43620480d0dd\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.391080 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58dc37ce-d8bf-46f0-9338-43620480d0dd-kube-api-access-x85g8" (OuterVolumeSpecName: "kube-api-access-x85g8") pod "58dc37ce-d8bf-46f0-9338-43620480d0dd" (UID: "58dc37ce-d8bf-46f0-9338-43620480d0dd"). InnerVolumeSpecName "kube-api-access-x85g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:17:46 crc kubenswrapper[4990]: E1203 13:17:46.408399 4990 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key podName:58dc37ce-d8bf-46f0-9338-43620480d0dd nodeName:}" failed. No retries permitted until 2025-12-03 13:17:46.90835577 +0000 UTC m=+2415.050266999 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key") pod "58dc37ce-d8bf-46f0-9338-43620480d0dd" (UID: "58dc37ce-d8bf-46f0-9338-43620480d0dd") : error deleting /var/lib/kubelet/pods/58dc37ce-d8bf-46f0-9338-43620480d0dd/volume-subpaths: remove /var/lib/kubelet/pods/58dc37ce-d8bf-46f0-9338-43620480d0dd/volume-subpaths: no such file or directory Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.411022 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-inventory" (OuterVolumeSpecName: "inventory") pod "58dc37ce-d8bf-46f0-9338-43620480d0dd" (UID: "58dc37ce-d8bf-46f0-9338-43620480d0dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.490044 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.490079 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x85g8\" (UniqueName: \"kubernetes.io/projected/58dc37ce-d8bf-46f0-9338-43620480d0dd-kube-api-access-x85g8\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.925697 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" event={"ID":"58dc37ce-d8bf-46f0-9338-43620480d0dd","Type":"ContainerDied","Data":"f44291a46cef18acdb2a7b2e5088f0a28891df0afac31ae96557e525e9732b94"} Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.925743 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f44291a46cef18acdb2a7b2e5088f0a28891df0afac31ae96557e525e9732b94" Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.925787 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hbnrv" Dec 03 13:17:46 crc kubenswrapper[4990]: I1203 13:17:46.999124 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key\") pod \"58dc37ce-d8bf-46f0-9338-43620480d0dd\" (UID: \"58dc37ce-d8bf-46f0-9338-43620480d0dd\") " Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.002974 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58dc37ce-d8bf-46f0-9338-43620480d0dd" (UID: "58dc37ce-d8bf-46f0-9338-43620480d0dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.017633 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p"] Dec 03 13:17:47 crc kubenswrapper[4990]: E1203 13:17:47.018187 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58dc37ce-d8bf-46f0-9338-43620480d0dd" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.018208 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="58dc37ce-d8bf-46f0-9338-43620480d0dd" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.018411 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="58dc37ce-d8bf-46f0-9338-43620480d0dd" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.019076 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.028773 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p"] Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.101637 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.101960 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.102212 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8thn7\" (UniqueName: \"kubernetes.io/projected/5398edd4-615e-4e57-b25b-e7732ef5dbf3-kube-api-access-8thn7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.102391 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58dc37ce-d8bf-46f0-9338-43620480d0dd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.204955 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.205109 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8thn7\" (UniqueName: \"kubernetes.io/projected/5398edd4-615e-4e57-b25b-e7732ef5dbf3-kube-api-access-8thn7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.205155 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.222711 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.222741 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8thn7\" (UniqueName: \"kubernetes.io/projected/5398edd4-615e-4e57-b25b-e7732ef5dbf3-kube-api-access-8thn7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.222828 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.375762 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.903201 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p"] Dec 03 13:17:47 crc kubenswrapper[4990]: W1203 13:17:47.908543 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5398edd4_615e_4e57_b25b_e7732ef5dbf3.slice/crio-bb0e94650bde72e7ce391efd6639c6be4a6126d12593085009b012f90946fbcb WatchSource:0}: Error finding container bb0e94650bde72e7ce391efd6639c6be4a6126d12593085009b012f90946fbcb: Status 404 returned error can't find the container with id bb0e94650bde72e7ce391efd6639c6be4a6126d12593085009b012f90946fbcb Dec 03 13:17:47 crc kubenswrapper[4990]: I1203 13:17:47.935130 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" event={"ID":"5398edd4-615e-4e57-b25b-e7732ef5dbf3","Type":"ContainerStarted","Data":"bb0e94650bde72e7ce391efd6639c6be4a6126d12593085009b012f90946fbcb"} Dec 03 13:17:51 crc kubenswrapper[4990]: I1203 13:17:51.977369 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" event={"ID":"5398edd4-615e-4e57-b25b-e7732ef5dbf3","Type":"ContainerStarted","Data":"7f7d43937414fee3e86976d2b3f849c26aa79cc75c18fda8f6e0bf72763f213e"} Dec 03 13:17:52 crc kubenswrapper[4990]: I1203 13:17:52.006569 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" podStartSLOduration=3.118982499 podStartE2EDuration="6.006544412s" podCreationTimestamp="2025-12-03 13:17:46 +0000 UTC" firstStartedPulling="2025-12-03 13:17:47.910842801 +0000 UTC m=+2416.052754030" lastFinishedPulling="2025-12-03 13:17:50.798404704 +0000 UTC m=+2418.940315943" observedRunningTime="2025-12-03 13:17:51.997062247 +0000 UTC m=+2420.138973516" watchObservedRunningTime="2025-12-03 13:17:52.006544412 +0000 UTC m=+2420.148455641" Dec 03 13:17:57 crc kubenswrapper[4990]: I1203 13:17:57.264679 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:17:57 crc kubenswrapper[4990]: E1203 13:17:57.265496 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:18:02 crc kubenswrapper[4990]: I1203 13:18:02.073900 4990 generic.go:334] "Generic (PLEG): container finished" podID="5398edd4-615e-4e57-b25b-e7732ef5dbf3" containerID="7f7d43937414fee3e86976d2b3f849c26aa79cc75c18fda8f6e0bf72763f213e" exitCode=0 Dec 03 13:18:02 crc kubenswrapper[4990]: I1203 13:18:02.074046 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" event={"ID":"5398edd4-615e-4e57-b25b-e7732ef5dbf3","Type":"ContainerDied","Data":"7f7d43937414fee3e86976d2b3f849c26aa79cc75c18fda8f6e0bf72763f213e"} Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.455920 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.550027 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-ssh-key\") pod \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.550119 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-inventory\") pod \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.550199 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8thn7\" (UniqueName: \"kubernetes.io/projected/5398edd4-615e-4e57-b25b-e7732ef5dbf3-kube-api-access-8thn7\") pod \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\" (UID: \"5398edd4-615e-4e57-b25b-e7732ef5dbf3\") " Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.555283 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5398edd4-615e-4e57-b25b-e7732ef5dbf3-kube-api-access-8thn7" (OuterVolumeSpecName: "kube-api-access-8thn7") pod "5398edd4-615e-4e57-b25b-e7732ef5dbf3" (UID: "5398edd4-615e-4e57-b25b-e7732ef5dbf3"). InnerVolumeSpecName "kube-api-access-8thn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.575845 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-inventory" (OuterVolumeSpecName: "inventory") pod "5398edd4-615e-4e57-b25b-e7732ef5dbf3" (UID: "5398edd4-615e-4e57-b25b-e7732ef5dbf3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.577383 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5398edd4-615e-4e57-b25b-e7732ef5dbf3" (UID: "5398edd4-615e-4e57-b25b-e7732ef5dbf3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.651660 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.651706 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5398edd4-615e-4e57-b25b-e7732ef5dbf3-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:03 crc kubenswrapper[4990]: I1203 13:18:03.651720 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8thn7\" (UniqueName: \"kubernetes.io/projected/5398edd4-615e-4e57-b25b-e7732ef5dbf3-kube-api-access-8thn7\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.091424 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" event={"ID":"5398edd4-615e-4e57-b25b-e7732ef5dbf3","Type":"ContainerDied","Data":"bb0e94650bde72e7ce391efd6639c6be4a6126d12593085009b012f90946fbcb"} Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.091828 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb0e94650bde72e7ce391efd6639c6be4a6126d12593085009b012f90946fbcb" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.091491 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.198460 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v"] Dec 03 13:18:04 crc kubenswrapper[4990]: E1203 13:18:04.198907 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5398edd4-615e-4e57-b25b-e7732ef5dbf3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.198924 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="5398edd4-615e-4e57-b25b-e7732ef5dbf3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.199112 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="5398edd4-615e-4e57-b25b-e7732ef5dbf3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.199779 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.201735 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.201919 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.202094 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.202276 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.202527 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.203532 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.203677 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.207879 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.225699 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v"] Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263713 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263760 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263794 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263814 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263848 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263876 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r557k\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-kube-api-access-r557k\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263897 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263915 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263936 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263953 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263969 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.263992 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.264008 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.264029 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365240 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365291 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r557k\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-kube-api-access-r557k\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365317 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365341 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365364 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365382 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365431 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365509 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365532 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365562 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365647 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365673 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365710 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.365734 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.370934 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.371064 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.371235 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.371477 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.371900 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.372408 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.372723 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.373250 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.373459 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.373948 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.374089 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.376196 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.377238 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.385862 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r557k\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-kube-api-access-r557k\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:04 crc kubenswrapper[4990]: I1203 13:18:04.515175 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:05 crc kubenswrapper[4990]: I1203 13:18:05.050421 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v"] Dec 03 13:18:05 crc kubenswrapper[4990]: I1203 13:18:05.101098 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" event={"ID":"213d788e-4e19-4020-a8ef-1d84d287fb95","Type":"ContainerStarted","Data":"63a43915f2197b55ab6661641af9ffe827455119ee62f2e0f303495c64bce07b"} Dec 03 13:18:06 crc kubenswrapper[4990]: I1203 13:18:06.114771 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" event={"ID":"213d788e-4e19-4020-a8ef-1d84d287fb95","Type":"ContainerStarted","Data":"b1dcac77e9db207bf415046a35b621d1b2f8ff2cbf14a108fc6aa5757d8d2828"} Dec 03 13:18:06 crc kubenswrapper[4990]: I1203 13:18:06.161014 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" podStartSLOduration=1.678433598 podStartE2EDuration="2.160993508s" podCreationTimestamp="2025-12-03 13:18:04 +0000 UTC" firstStartedPulling="2025-12-03 13:18:05.057381524 +0000 UTC m=+2433.199292753" lastFinishedPulling="2025-12-03 13:18:05.539941434 +0000 UTC m=+2433.681852663" observedRunningTime="2025-12-03 13:18:06.140975361 +0000 UTC m=+2434.282886640" watchObservedRunningTime="2025-12-03 13:18:06.160993508 +0000 UTC m=+2434.302904747" Dec 03 13:18:11 crc kubenswrapper[4990]: I1203 13:18:11.264396 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:18:11 crc kubenswrapper[4990]: E1203 13:18:11.265130 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:18:26 crc kubenswrapper[4990]: I1203 13:18:26.265242 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:18:26 crc kubenswrapper[4990]: E1203 13:18:26.266009 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:18:40 crc kubenswrapper[4990]: I1203 13:18:40.265610 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:18:40 crc kubenswrapper[4990]: E1203 13:18:40.266680 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:18:44 crc kubenswrapper[4990]: I1203 13:18:44.491038 4990 generic.go:334] "Generic (PLEG): container finished" podID="213d788e-4e19-4020-a8ef-1d84d287fb95" containerID="b1dcac77e9db207bf415046a35b621d1b2f8ff2cbf14a108fc6aa5757d8d2828" exitCode=0 Dec 03 13:18:44 crc kubenswrapper[4990]: I1203 13:18:44.491128 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" event={"ID":"213d788e-4e19-4020-a8ef-1d84d287fb95","Type":"ContainerDied","Data":"b1dcac77e9db207bf415046a35b621d1b2f8ff2cbf14a108fc6aa5757d8d2828"} Dec 03 13:18:45 crc kubenswrapper[4990]: I1203 13:18:45.901087 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.005954 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-inventory\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007002 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007062 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-libvirt-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007175 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007216 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-telemetry-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007295 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ovn-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007347 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ssh-key\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007388 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-nova-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007434 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-repo-setup-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007508 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-bootstrap-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007545 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007626 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-ovn-default-certs-0\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007679 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-neutron-metadata-combined-ca-bundle\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.007710 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r557k\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-kube-api-access-r557k\") pod \"213d788e-4e19-4020-a8ef-1d84d287fb95\" (UID: \"213d788e-4e19-4020-a8ef-1d84d287fb95\") " Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.013979 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.014592 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.014818 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.014914 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.014971 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.015632 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.016148 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.017996 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.018012 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.018019 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.018108 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-kube-api-access-r557k" (OuterVolumeSpecName: "kube-api-access-r557k") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "kube-api-access-r557k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.019329 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.041849 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-inventory" (OuterVolumeSpecName: "inventory") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.045656 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "213d788e-4e19-4020-a8ef-1d84d287fb95" (UID: "213d788e-4e19-4020-a8ef-1d84d287fb95"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111135 4990 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111181 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111192 4990 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111203 4990 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111216 4990 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111233 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111247 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111260 4990 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111272 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r557k\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-kube-api-access-r557k\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111284 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111295 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111309 4990 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111321 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/213d788e-4e19-4020-a8ef-1d84d287fb95-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.111331 4990 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213d788e-4e19-4020-a8ef-1d84d287fb95-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.511564 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" event={"ID":"213d788e-4e19-4020-a8ef-1d84d287fb95","Type":"ContainerDied","Data":"63a43915f2197b55ab6661641af9ffe827455119ee62f2e0f303495c64bce07b"} Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.511614 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a43915f2197b55ab6661641af9ffe827455119ee62f2e0f303495c64bce07b" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.511616 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.596971 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t"] Dec 03 13:18:46 crc kubenswrapper[4990]: E1203 13:18:46.606236 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213d788e-4e19-4020-a8ef-1d84d287fb95" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.606562 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="213d788e-4e19-4020-a8ef-1d84d287fb95" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.607141 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="213d788e-4e19-4020-a8ef-1d84d287fb95" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.608326 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t"] Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.608542 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.612614 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.613094 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.613619 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.613993 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.616185 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.723175 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff2m4\" (UniqueName: \"kubernetes.io/projected/3cf5d887-0878-407b-98c9-539cb967a73e-kube-api-access-ff2m4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.723343 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.723380 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3cf5d887-0878-407b-98c9-539cb967a73e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.723442 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.723513 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.825732 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.826101 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3cf5d887-0878-407b-98c9-539cb967a73e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.826243 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.826341 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.826518 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff2m4\" (UniqueName: \"kubernetes.io/projected/3cf5d887-0878-407b-98c9-539cb967a73e-kube-api-access-ff2m4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.827195 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3cf5d887-0878-407b-98c9-539cb967a73e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.829611 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.830603 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.831319 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.844314 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff2m4\" (UniqueName: \"kubernetes.io/projected/3cf5d887-0878-407b-98c9-539cb967a73e-kube-api-access-ff2m4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q8c6t\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:46 crc kubenswrapper[4990]: I1203 13:18:46.930743 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:18:47 crc kubenswrapper[4990]: I1203 13:18:47.459080 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t"] Dec 03 13:18:47 crc kubenswrapper[4990]: W1203 13:18:47.464559 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cf5d887_0878_407b_98c9_539cb967a73e.slice/crio-a7db6481c6c476c175e8a025c8d5d00aaa995f6ad2c27fb1925cf253eade7c20 WatchSource:0}: Error finding container a7db6481c6c476c175e8a025c8d5d00aaa995f6ad2c27fb1925cf253eade7c20: Status 404 returned error can't find the container with id a7db6481c6c476c175e8a025c8d5d00aaa995f6ad2c27fb1925cf253eade7c20 Dec 03 13:18:47 crc kubenswrapper[4990]: I1203 13:18:47.522603 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" event={"ID":"3cf5d887-0878-407b-98c9-539cb967a73e","Type":"ContainerStarted","Data":"a7db6481c6c476c175e8a025c8d5d00aaa995f6ad2c27fb1925cf253eade7c20"} Dec 03 13:18:48 crc kubenswrapper[4990]: I1203 13:18:48.533729 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" event={"ID":"3cf5d887-0878-407b-98c9-539cb967a73e","Type":"ContainerStarted","Data":"d3da7abaaebb5feba1f38a532b1c1d30101d9f93c2d58812be6d20f25c247f8c"} Dec 03 13:18:48 crc kubenswrapper[4990]: I1203 13:18:48.557870 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" podStartSLOduration=2.113303837 podStartE2EDuration="2.557850757s" podCreationTimestamp="2025-12-03 13:18:46 +0000 UTC" firstStartedPulling="2025-12-03 13:18:47.46719169 +0000 UTC m=+2475.609102919" lastFinishedPulling="2025-12-03 13:18:47.91173861 +0000 UTC m=+2476.053649839" observedRunningTime="2025-12-03 13:18:48.550592662 +0000 UTC m=+2476.692503891" watchObservedRunningTime="2025-12-03 13:18:48.557850757 +0000 UTC m=+2476.699761986" Dec 03 13:18:52 crc kubenswrapper[4990]: I1203 13:18:52.271840 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:18:52 crc kubenswrapper[4990]: E1203 13:18:52.272433 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:19:04 crc kubenswrapper[4990]: I1203 13:19:04.264980 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:19:04 crc kubenswrapper[4990]: E1203 13:19:04.265839 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:19:15 crc kubenswrapper[4990]: I1203 13:19:15.264938 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:19:15 crc kubenswrapper[4990]: E1203 13:19:15.265752 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:19:28 crc kubenswrapper[4990]: I1203 13:19:28.264594 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:19:28 crc kubenswrapper[4990]: E1203 13:19:28.265459 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:19:43 crc kubenswrapper[4990]: I1203 13:19:43.264617 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:19:43 crc kubenswrapper[4990]: E1203 13:19:43.265491 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:19:53 crc kubenswrapper[4990]: I1203 13:19:53.156354 4990 generic.go:334] "Generic (PLEG): container finished" podID="3cf5d887-0878-407b-98c9-539cb967a73e" containerID="d3da7abaaebb5feba1f38a532b1c1d30101d9f93c2d58812be6d20f25c247f8c" exitCode=0 Dec 03 13:19:53 crc kubenswrapper[4990]: I1203 13:19:53.156443 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" event={"ID":"3cf5d887-0878-407b-98c9-539cb967a73e","Type":"ContainerDied","Data":"d3da7abaaebb5feba1f38a532b1c1d30101d9f93c2d58812be6d20f25c247f8c"} Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.264973 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:19:54 crc kubenswrapper[4990]: E1203 13:19:54.265511 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.558574 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.585519 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ovn-combined-ca-bundle\") pod \"3cf5d887-0878-407b-98c9-539cb967a73e\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.585682 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ssh-key\") pod \"3cf5d887-0878-407b-98c9-539cb967a73e\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.585761 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-inventory\") pod \"3cf5d887-0878-407b-98c9-539cb967a73e\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.585788 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff2m4\" (UniqueName: \"kubernetes.io/projected/3cf5d887-0878-407b-98c9-539cb967a73e-kube-api-access-ff2m4\") pod \"3cf5d887-0878-407b-98c9-539cb967a73e\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.585881 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3cf5d887-0878-407b-98c9-539cb967a73e-ovncontroller-config-0\") pod \"3cf5d887-0878-407b-98c9-539cb967a73e\" (UID: \"3cf5d887-0878-407b-98c9-539cb967a73e\") " Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.604111 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3cf5d887-0878-407b-98c9-539cb967a73e" (UID: "3cf5d887-0878-407b-98c9-539cb967a73e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.604261 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf5d887-0878-407b-98c9-539cb967a73e-kube-api-access-ff2m4" (OuterVolumeSpecName: "kube-api-access-ff2m4") pod "3cf5d887-0878-407b-98c9-539cb967a73e" (UID: "3cf5d887-0878-407b-98c9-539cb967a73e"). InnerVolumeSpecName "kube-api-access-ff2m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.614957 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3cf5d887-0878-407b-98c9-539cb967a73e" (UID: "3cf5d887-0878-407b-98c9-539cb967a73e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.615081 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-inventory" (OuterVolumeSpecName: "inventory") pod "3cf5d887-0878-407b-98c9-539cb967a73e" (UID: "3cf5d887-0878-407b-98c9-539cb967a73e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.619658 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf5d887-0878-407b-98c9-539cb967a73e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3cf5d887-0878-407b-98c9-539cb967a73e" (UID: "3cf5d887-0878-407b-98c9-539cb967a73e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.688315 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.688355 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.688371 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff2m4\" (UniqueName: \"kubernetes.io/projected/3cf5d887-0878-407b-98c9-539cb967a73e-kube-api-access-ff2m4\") on node \"crc\" DevicePath \"\"" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.688387 4990 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3cf5d887-0878-407b-98c9-539cb967a73e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:19:54 crc kubenswrapper[4990]: I1203 13:19:54.688399 4990 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf5d887-0878-407b-98c9-539cb967a73e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.172022 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" event={"ID":"3cf5d887-0878-407b-98c9-539cb967a73e","Type":"ContainerDied","Data":"a7db6481c6c476c175e8a025c8d5d00aaa995f6ad2c27fb1925cf253eade7c20"} Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.172117 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7db6481c6c476c175e8a025c8d5d00aaa995f6ad2c27fb1925cf253eade7c20" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.172234 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q8c6t" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.328714 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w"] Dec 03 13:19:55 crc kubenswrapper[4990]: E1203 13:19:55.329382 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf5d887-0878-407b-98c9-539cb967a73e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.329403 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf5d887-0878-407b-98c9-539cb967a73e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.329700 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf5d887-0878-407b-98c9-539cb967a73e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.330571 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.334075 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.334105 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.334170 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.334077 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.334258 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.334301 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.341797 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w"] Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.400596 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.400918 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpzkh\" (UniqueName: \"kubernetes.io/projected/52371a09-a7eb-42bb-aa53-3a2188c6658d-kube-api-access-jpzkh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.401034 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.401167 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.401301 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.401525 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.503821 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.504214 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.504267 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpzkh\" (UniqueName: \"kubernetes.io/projected/52371a09-a7eb-42bb-aa53-3a2188c6658d-kube-api-access-jpzkh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.504310 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.504349 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.504392 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.508148 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.508700 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.509021 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.510240 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.511665 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.521556 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpzkh\" (UniqueName: \"kubernetes.io/projected/52371a09-a7eb-42bb-aa53-3a2188c6658d-kube-api-access-jpzkh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:55 crc kubenswrapper[4990]: I1203 13:19:55.653902 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:19:56 crc kubenswrapper[4990]: I1203 13:19:56.205756 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w"] Dec 03 13:19:56 crc kubenswrapper[4990]: I1203 13:19:56.213260 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:19:57 crc kubenswrapper[4990]: I1203 13:19:57.201328 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" event={"ID":"52371a09-a7eb-42bb-aa53-3a2188c6658d","Type":"ContainerStarted","Data":"44e538ffaa20b56718c786499744afaffa37797f707cf445bdd7565a9325a59b"} Dec 03 13:19:57 crc kubenswrapper[4990]: I1203 13:19:57.201788 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" event={"ID":"52371a09-a7eb-42bb-aa53-3a2188c6658d","Type":"ContainerStarted","Data":"d950a11261b447331c2cf76cc86bd66723d41b0b812ad842efb828a2c71922c9"} Dec 03 13:19:57 crc kubenswrapper[4990]: I1203 13:19:57.223692 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" podStartSLOduration=1.6853546000000001 podStartE2EDuration="2.223672245s" podCreationTimestamp="2025-12-03 13:19:55 +0000 UTC" firstStartedPulling="2025-12-03 13:19:56.213079485 +0000 UTC m=+2544.354990714" lastFinishedPulling="2025-12-03 13:19:56.75139713 +0000 UTC m=+2544.893308359" observedRunningTime="2025-12-03 13:19:57.217429157 +0000 UTC m=+2545.359340386" watchObservedRunningTime="2025-12-03 13:19:57.223672245 +0000 UTC m=+2545.365583474" Dec 03 13:20:07 crc kubenswrapper[4990]: I1203 13:20:07.264430 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:20:08 crc kubenswrapper[4990]: I1203 13:20:08.302343 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"8e94b8c8d37ddcf489349bd8393b94b78192224584ef611211b491c2bcda4d09"} Dec 03 13:20:43 crc kubenswrapper[4990]: I1203 13:20:43.665728 4990 generic.go:334] "Generic (PLEG): container finished" podID="52371a09-a7eb-42bb-aa53-3a2188c6658d" containerID="44e538ffaa20b56718c786499744afaffa37797f707cf445bdd7565a9325a59b" exitCode=0 Dec 03 13:20:43 crc kubenswrapper[4990]: I1203 13:20:43.665939 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" event={"ID":"52371a09-a7eb-42bb-aa53-3a2188c6658d","Type":"ContainerDied","Data":"44e538ffaa20b56718c786499744afaffa37797f707cf445bdd7565a9325a59b"} Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.109276 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.197377 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-ssh-key\") pod \"52371a09-a7eb-42bb-aa53-3a2188c6658d\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.197498 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpzkh\" (UniqueName: \"kubernetes.io/projected/52371a09-a7eb-42bb-aa53-3a2188c6658d-kube-api-access-jpzkh\") pod \"52371a09-a7eb-42bb-aa53-3a2188c6658d\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.197535 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-nova-metadata-neutron-config-0\") pod \"52371a09-a7eb-42bb-aa53-3a2188c6658d\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.197572 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-inventory\") pod \"52371a09-a7eb-42bb-aa53-3a2188c6658d\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.197706 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-metadata-combined-ca-bundle\") pod \"52371a09-a7eb-42bb-aa53-3a2188c6658d\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.197779 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"52371a09-a7eb-42bb-aa53-3a2188c6658d\" (UID: \"52371a09-a7eb-42bb-aa53-3a2188c6658d\") " Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.204652 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52371a09-a7eb-42bb-aa53-3a2188c6658d-kube-api-access-jpzkh" (OuterVolumeSpecName: "kube-api-access-jpzkh") pod "52371a09-a7eb-42bb-aa53-3a2188c6658d" (UID: "52371a09-a7eb-42bb-aa53-3a2188c6658d"). InnerVolumeSpecName "kube-api-access-jpzkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.205545 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "52371a09-a7eb-42bb-aa53-3a2188c6658d" (UID: "52371a09-a7eb-42bb-aa53-3a2188c6658d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.229303 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "52371a09-a7eb-42bb-aa53-3a2188c6658d" (UID: "52371a09-a7eb-42bb-aa53-3a2188c6658d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.237831 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-inventory" (OuterVolumeSpecName: "inventory") pod "52371a09-a7eb-42bb-aa53-3a2188c6658d" (UID: "52371a09-a7eb-42bb-aa53-3a2188c6658d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.241107 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "52371a09-a7eb-42bb-aa53-3a2188c6658d" (UID: "52371a09-a7eb-42bb-aa53-3a2188c6658d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.258631 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "52371a09-a7eb-42bb-aa53-3a2188c6658d" (UID: "52371a09-a7eb-42bb-aa53-3a2188c6658d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.300133 4990 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.302725 4990 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.302865 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.302937 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpzkh\" (UniqueName: \"kubernetes.io/projected/52371a09-a7eb-42bb-aa53-3a2188c6658d-kube-api-access-jpzkh\") on node \"crc\" DevicePath \"\"" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.303017 4990 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.303084 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52371a09-a7eb-42bb-aa53-3a2188c6658d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.687501 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" event={"ID":"52371a09-a7eb-42bb-aa53-3a2188c6658d","Type":"ContainerDied","Data":"d950a11261b447331c2cf76cc86bd66723d41b0b812ad842efb828a2c71922c9"} Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.687557 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d950a11261b447331c2cf76cc86bd66723d41b0b812ad842efb828a2c71922c9" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.687577 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.804147 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk"] Dec 03 13:20:45 crc kubenswrapper[4990]: E1203 13:20:45.804766 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52371a09-a7eb-42bb-aa53-3a2188c6658d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.804843 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="52371a09-a7eb-42bb-aa53-3a2188c6658d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.805091 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="52371a09-a7eb-42bb-aa53-3a2188c6658d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.806260 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.809259 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.809717 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.809887 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.810144 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.810254 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.816776 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk"] Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.914333 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.914532 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.914557 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.914638 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:45 crc kubenswrapper[4990]: I1203 13:20:45.914690 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x65sp\" (UniqueName: \"kubernetes.io/projected/4c4068fd-3e0a-442a-b035-f2e84fba7e88-kube-api-access-x65sp\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.017240 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.017400 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x65sp\" (UniqueName: \"kubernetes.io/projected/4c4068fd-3e0a-442a-b035-f2e84fba7e88-kube-api-access-x65sp\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.017560 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.017861 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.017922 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.021775 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.022705 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.027175 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.029926 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.038406 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x65sp\" (UniqueName: \"kubernetes.io/projected/4c4068fd-3e0a-442a-b035-f2e84fba7e88-kube-api-access-x65sp\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.136661 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:20:46 crc kubenswrapper[4990]: I1203 13:20:46.745037 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk"] Dec 03 13:20:47 crc kubenswrapper[4990]: I1203 13:20:47.712851 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" event={"ID":"4c4068fd-3e0a-442a-b035-f2e84fba7e88","Type":"ContainerStarted","Data":"cb6504cd1a238b9691e2e8d8e2c7cdf81b1627daeb626915849639e8cb001027"} Dec 03 13:20:48 crc kubenswrapper[4990]: I1203 13:20:48.728321 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" event={"ID":"4c4068fd-3e0a-442a-b035-f2e84fba7e88","Type":"ContainerStarted","Data":"f9644704032a81679f826adf38416c61ef8f424c15b16e29b812292796e6a338"} Dec 03 13:20:48 crc kubenswrapper[4990]: I1203 13:20:48.763580 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" podStartSLOduration=2.903493774 podStartE2EDuration="3.763539722s" podCreationTimestamp="2025-12-03 13:20:45 +0000 UTC" firstStartedPulling="2025-12-03 13:20:46.75907814 +0000 UTC m=+2594.900989379" lastFinishedPulling="2025-12-03 13:20:47.619124088 +0000 UTC m=+2595.761035327" observedRunningTime="2025-12-03 13:20:48.749498374 +0000 UTC m=+2596.891409613" watchObservedRunningTime="2025-12-03 13:20:48.763539722 +0000 UTC m=+2596.905450961" Dec 03 13:22:33 crc kubenswrapper[4990]: I1203 13:22:33.287097 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:22:33 crc kubenswrapper[4990]: I1203 13:22:33.287980 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:23:03 crc kubenswrapper[4990]: I1203 13:23:03.301134 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:23:03 crc kubenswrapper[4990]: I1203 13:23:03.302051 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.287091 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.287795 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.287847 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.288656 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e94b8c8d37ddcf489349bd8393b94b78192224584ef611211b491c2bcda4d09"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.288715 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://8e94b8c8d37ddcf489349bd8393b94b78192224584ef611211b491c2bcda4d09" gracePeriod=600 Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.744779 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="8e94b8c8d37ddcf489349bd8393b94b78192224584ef611211b491c2bcda4d09" exitCode=0 Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.744840 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"8e94b8c8d37ddcf489349bd8393b94b78192224584ef611211b491c2bcda4d09"} Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.745534 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5"} Dec 03 13:23:33 crc kubenswrapper[4990]: I1203 13:23:33.745571 4990 scope.go:117] "RemoveContainer" containerID="31a287d57cde121911ebc9693e8d442b111dd09dad294266a69836869272da4b" Dec 03 13:25:25 crc kubenswrapper[4990]: I1203 13:25:25.832819 4990 generic.go:334] "Generic (PLEG): container finished" podID="4c4068fd-3e0a-442a-b035-f2e84fba7e88" containerID="f9644704032a81679f826adf38416c61ef8f424c15b16e29b812292796e6a338" exitCode=0 Dec 03 13:25:25 crc kubenswrapper[4990]: I1203 13:25:25.832903 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" event={"ID":"4c4068fd-3e0a-442a-b035-f2e84fba7e88","Type":"ContainerDied","Data":"f9644704032a81679f826adf38416c61ef8f424c15b16e29b812292796e6a338"} Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.266126 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.425249 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-ssh-key\") pod \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.425410 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x65sp\" (UniqueName: \"kubernetes.io/projected/4c4068fd-3e0a-442a-b035-f2e84fba7e88-kube-api-access-x65sp\") pod \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.425510 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-inventory\") pod \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.425647 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-combined-ca-bundle\") pod \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.425683 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-secret-0\") pod \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\" (UID: \"4c4068fd-3e0a-442a-b035-f2e84fba7e88\") " Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.433260 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c4068fd-3e0a-442a-b035-f2e84fba7e88-kube-api-access-x65sp" (OuterVolumeSpecName: "kube-api-access-x65sp") pod "4c4068fd-3e0a-442a-b035-f2e84fba7e88" (UID: "4c4068fd-3e0a-442a-b035-f2e84fba7e88"). InnerVolumeSpecName "kube-api-access-x65sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.436726 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4c4068fd-3e0a-442a-b035-f2e84fba7e88" (UID: "4c4068fd-3e0a-442a-b035-f2e84fba7e88"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.463959 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4c4068fd-3e0a-442a-b035-f2e84fba7e88" (UID: "4c4068fd-3e0a-442a-b035-f2e84fba7e88"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.467224 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-inventory" (OuterVolumeSpecName: "inventory") pod "4c4068fd-3e0a-442a-b035-f2e84fba7e88" (UID: "4c4068fd-3e0a-442a-b035-f2e84fba7e88"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.472295 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4c4068fd-3e0a-442a-b035-f2e84fba7e88" (UID: "4c4068fd-3e0a-442a-b035-f2e84fba7e88"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.528964 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.529210 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x65sp\" (UniqueName: \"kubernetes.io/projected/4c4068fd-3e0a-442a-b035-f2e84fba7e88-kube-api-access-x65sp\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.529228 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.529241 4990 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.529253 4990 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c4068fd-3e0a-442a-b035-f2e84fba7e88-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.853430 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" event={"ID":"4c4068fd-3e0a-442a-b035-f2e84fba7e88","Type":"ContainerDied","Data":"cb6504cd1a238b9691e2e8d8e2c7cdf81b1627daeb626915849639e8cb001027"} Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.853494 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb6504cd1a238b9691e2e8d8e2c7cdf81b1627daeb626915849639e8cb001027" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.853527 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.968905 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts"] Dec 03 13:25:27 crc kubenswrapper[4990]: E1203 13:25:27.969921 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c4068fd-3e0a-442a-b035-f2e84fba7e88" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.969943 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c4068fd-3e0a-442a-b035-f2e84fba7e88" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.970195 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c4068fd-3e0a-442a-b035-f2e84fba7e88" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.970866 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.973214 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.973276 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.973276 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.973369 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.973591 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.973845 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.974866 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:25:27 crc kubenswrapper[4990]: I1203 13:25:27.990545 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts"] Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038220 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038278 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038349 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4191ee3d-7722-4690-938c-7b8d8478589e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038383 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b72v6\" (UniqueName: \"kubernetes.io/projected/4191ee3d-7722-4690-938c-7b8d8478589e-kube-api-access-b72v6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038566 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038585 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038612 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038636 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.038651 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.140563 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4191ee3d-7722-4690-938c-7b8d8478589e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.140656 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b72v6\" (UniqueName: \"kubernetes.io/projected/4191ee3d-7722-4690-938c-7b8d8478589e-kube-api-access-b72v6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141081 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141164 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141225 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141268 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141303 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141384 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.141418 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.142391 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4191ee3d-7722-4690-938c-7b8d8478589e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.144880 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.146250 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.146348 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.146807 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.147729 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.149254 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.149796 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.158944 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b72v6\" (UniqueName: \"kubernetes.io/projected/4191ee3d-7722-4690-938c-7b8d8478589e-kube-api-access-b72v6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hhvts\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.289519 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.920073 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts"] Dec 03 13:25:28 crc kubenswrapper[4990]: I1203 13:25:28.926699 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:25:29 crc kubenswrapper[4990]: I1203 13:25:29.872253 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" event={"ID":"4191ee3d-7722-4690-938c-7b8d8478589e","Type":"ContainerStarted","Data":"048dd83e63c4f1ac2c8da86ec3401d92a31a3c9d957d8a01536389b431972d24"} Dec 03 13:25:29 crc kubenswrapper[4990]: I1203 13:25:29.872735 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" event={"ID":"4191ee3d-7722-4690-938c-7b8d8478589e","Type":"ContainerStarted","Data":"58a4fc7ce39faf10ef75ba642911d151c8c151722fb53fe487ee38fbe17bdb0e"} Dec 03 13:25:29 crc kubenswrapper[4990]: I1203 13:25:29.894998 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" podStartSLOduration=2.375454273 podStartE2EDuration="2.894977171s" podCreationTimestamp="2025-12-03 13:25:27 +0000 UTC" firstStartedPulling="2025-12-03 13:25:28.926385967 +0000 UTC m=+2877.068297196" lastFinishedPulling="2025-12-03 13:25:29.445908855 +0000 UTC m=+2877.587820094" observedRunningTime="2025-12-03 13:25:29.889036564 +0000 UTC m=+2878.030947803" watchObservedRunningTime="2025-12-03 13:25:29.894977171 +0000 UTC m=+2878.036888400" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.226252 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrb2f"] Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.229765 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.236877 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrb2f"] Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.321499 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c2sd\" (UniqueName: \"kubernetes.io/projected/bb300e54-c925-4f74-97d5-8a189217b86b-kube-api-access-7c2sd\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.321546 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-catalog-content\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.321569 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-utilities\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.424379 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c2sd\" (UniqueName: \"kubernetes.io/projected/bb300e54-c925-4f74-97d5-8a189217b86b-kube-api-access-7c2sd\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.424432 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-catalog-content\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.424487 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-utilities\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.424994 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-catalog-content\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.425132 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-utilities\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.444127 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c2sd\" (UniqueName: \"kubernetes.io/projected/bb300e54-c925-4f74-97d5-8a189217b86b-kube-api-access-7c2sd\") pod \"redhat-operators-lrb2f\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:32 crc kubenswrapper[4990]: I1203 13:25:32.564233 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:33 crc kubenswrapper[4990]: I1203 13:25:33.058040 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrb2f"] Dec 03 13:25:33 crc kubenswrapper[4990]: I1203 13:25:33.286241 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:25:33 crc kubenswrapper[4990]: I1203 13:25:33.286730 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:25:33 crc kubenswrapper[4990]: I1203 13:25:33.907579 4990 generic.go:334] "Generic (PLEG): container finished" podID="bb300e54-c925-4f74-97d5-8a189217b86b" containerID="b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad" exitCode=0 Dec 03 13:25:33 crc kubenswrapper[4990]: I1203 13:25:33.907644 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerDied","Data":"b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad"} Dec 03 13:25:33 crc kubenswrapper[4990]: I1203 13:25:33.907673 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerStarted","Data":"6b66b882e7861334b56bac12735bb8074d61c9b806160387d1b87c93cfd94f67"} Dec 03 13:25:35 crc kubenswrapper[4990]: I1203 13:25:35.947271 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerStarted","Data":"83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12"} Dec 03 13:25:38 crc kubenswrapper[4990]: I1203 13:25:38.977546 4990 generic.go:334] "Generic (PLEG): container finished" podID="bb300e54-c925-4f74-97d5-8a189217b86b" containerID="83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12" exitCode=0 Dec 03 13:25:38 crc kubenswrapper[4990]: I1203 13:25:38.977631 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerDied","Data":"83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12"} Dec 03 13:25:39 crc kubenswrapper[4990]: I1203 13:25:39.993000 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerStarted","Data":"28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62"} Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.014937 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrb2f" podStartSLOduration=2.213958482 podStartE2EDuration="8.014917134s" podCreationTimestamp="2025-12-03 13:25:32 +0000 UTC" firstStartedPulling="2025-12-03 13:25:33.909394578 +0000 UTC m=+2882.051305807" lastFinishedPulling="2025-12-03 13:25:39.71035323 +0000 UTC m=+2887.852264459" observedRunningTime="2025-12-03 13:25:40.010562619 +0000 UTC m=+2888.152473868" watchObservedRunningTime="2025-12-03 13:25:40.014917134 +0000 UTC m=+2888.156828363" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.034812 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vrnhf"] Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.037416 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.070708 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrnhf"] Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.184500 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-utilities\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.184574 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8fqz\" (UniqueName: \"kubernetes.io/projected/a8619cf9-e617-41e4-8c91-56fbd9c71f20-kube-api-access-x8fqz\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.184757 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-catalog-content\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.287045 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-utilities\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.287116 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8fqz\" (UniqueName: \"kubernetes.io/projected/a8619cf9-e617-41e4-8c91-56fbd9c71f20-kube-api-access-x8fqz\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.287179 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-catalog-content\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.287804 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-utilities\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.287851 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-catalog-content\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.315446 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8fqz\" (UniqueName: \"kubernetes.io/projected/a8619cf9-e617-41e4-8c91-56fbd9c71f20-kube-api-access-x8fqz\") pod \"community-operators-vrnhf\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:40 crc kubenswrapper[4990]: I1203 13:25:40.367937 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:41 crc kubenswrapper[4990]: I1203 13:25:41.019356 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrnhf"] Dec 03 13:25:42 crc kubenswrapper[4990]: I1203 13:25:42.019853 4990 generic.go:334] "Generic (PLEG): container finished" podID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerID="c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7" exitCode=0 Dec 03 13:25:42 crc kubenswrapper[4990]: I1203 13:25:42.019984 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrnhf" event={"ID":"a8619cf9-e617-41e4-8c91-56fbd9c71f20","Type":"ContainerDied","Data":"c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7"} Dec 03 13:25:42 crc kubenswrapper[4990]: I1203 13:25:42.020269 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrnhf" event={"ID":"a8619cf9-e617-41e4-8c91-56fbd9c71f20","Type":"ContainerStarted","Data":"849bcfe054edd42a86af32efef01495a6327ebb4691140491655c23ef311e3ec"} Dec 03 13:25:42 crc kubenswrapper[4990]: I1203 13:25:42.565321 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:42 crc kubenswrapper[4990]: I1203 13:25:42.565394 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:43 crc kubenswrapper[4990]: I1203 13:25:43.611012 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrb2f" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="registry-server" probeResult="failure" output=< Dec 03 13:25:43 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 13:25:43 crc kubenswrapper[4990]: > Dec 03 13:25:44 crc kubenswrapper[4990]: I1203 13:25:44.038718 4990 generic.go:334] "Generic (PLEG): container finished" podID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerID="7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a" exitCode=0 Dec 03 13:25:44 crc kubenswrapper[4990]: I1203 13:25:44.038820 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrnhf" event={"ID":"a8619cf9-e617-41e4-8c91-56fbd9c71f20","Type":"ContainerDied","Data":"7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a"} Dec 03 13:25:45 crc kubenswrapper[4990]: I1203 13:25:45.050667 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrnhf" event={"ID":"a8619cf9-e617-41e4-8c91-56fbd9c71f20","Type":"ContainerStarted","Data":"69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9"} Dec 03 13:25:45 crc kubenswrapper[4990]: I1203 13:25:45.074075 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vrnhf" podStartSLOduration=2.644396048 podStartE2EDuration="5.074052468s" podCreationTimestamp="2025-12-03 13:25:40 +0000 UTC" firstStartedPulling="2025-12-03 13:25:42.022352449 +0000 UTC m=+2890.164263678" lastFinishedPulling="2025-12-03 13:25:44.452008869 +0000 UTC m=+2892.593920098" observedRunningTime="2025-12-03 13:25:45.068125642 +0000 UTC m=+2893.210036881" watchObservedRunningTime="2025-12-03 13:25:45.074052468 +0000 UTC m=+2893.215963697" Dec 03 13:25:50 crc kubenswrapper[4990]: I1203 13:25:50.368375 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:50 crc kubenswrapper[4990]: I1203 13:25:50.369107 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:50 crc kubenswrapper[4990]: I1203 13:25:50.416834 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:51 crc kubenswrapper[4990]: I1203 13:25:51.156901 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:51 crc kubenswrapper[4990]: I1203 13:25:51.203964 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrnhf"] Dec 03 13:25:52 crc kubenswrapper[4990]: I1203 13:25:52.609971 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:52 crc kubenswrapper[4990]: I1203 13:25:52.655591 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:25:53 crc kubenswrapper[4990]: I1203 13:25:53.120977 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vrnhf" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="registry-server" containerID="cri-o://69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9" gracePeriod=2 Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.060257 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.094664 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-catalog-content\") pod \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.094832 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-utilities\") pod \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.094856 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8fqz\" (UniqueName: \"kubernetes.io/projected/a8619cf9-e617-41e4-8c91-56fbd9c71f20-kube-api-access-x8fqz\") pod \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\" (UID: \"a8619cf9-e617-41e4-8c91-56fbd9c71f20\") " Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.095696 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-utilities" (OuterVolumeSpecName: "utilities") pod "a8619cf9-e617-41e4-8c91-56fbd9c71f20" (UID: "a8619cf9-e617-41e4-8c91-56fbd9c71f20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.100796 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8619cf9-e617-41e4-8c91-56fbd9c71f20-kube-api-access-x8fqz" (OuterVolumeSpecName: "kube-api-access-x8fqz") pod "a8619cf9-e617-41e4-8c91-56fbd9c71f20" (UID: "a8619cf9-e617-41e4-8c91-56fbd9c71f20"). InnerVolumeSpecName "kube-api-access-x8fqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.148915 4990 generic.go:334] "Generic (PLEG): container finished" podID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerID="69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9" exitCode=0 Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.148995 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrnhf" event={"ID":"a8619cf9-e617-41e4-8c91-56fbd9c71f20","Type":"ContainerDied","Data":"69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9"} Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.149021 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrnhf" event={"ID":"a8619cf9-e617-41e4-8c91-56fbd9c71f20","Type":"ContainerDied","Data":"849bcfe054edd42a86af32efef01495a6327ebb4691140491655c23ef311e3ec"} Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.149022 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrnhf" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.149037 4990 scope.go:117] "RemoveContainer" containerID="69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.149842 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8619cf9-e617-41e4-8c91-56fbd9c71f20" (UID: "a8619cf9-e617-41e4-8c91-56fbd9c71f20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.171856 4990 scope.go:117] "RemoveContainer" containerID="7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.196922 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8fqz\" (UniqueName: \"kubernetes.io/projected/a8619cf9-e617-41e4-8c91-56fbd9c71f20-kube-api-access-x8fqz\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.196959 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.197000 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8619cf9-e617-41e4-8c91-56fbd9c71f20-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.203424 4990 scope.go:117] "RemoveContainer" containerID="c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.248867 4990 scope.go:117] "RemoveContainer" containerID="69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9" Dec 03 13:25:54 crc kubenswrapper[4990]: E1203 13:25:54.249235 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9\": container with ID starting with 69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9 not found: ID does not exist" containerID="69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.249273 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9"} err="failed to get container status \"69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9\": rpc error: code = NotFound desc = could not find container \"69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9\": container with ID starting with 69fcf31caf61533ed7579b6c26972c757280b9fb304fcca17e9550a7f2f345d9 not found: ID does not exist" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.249298 4990 scope.go:117] "RemoveContainer" containerID="7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a" Dec 03 13:25:54 crc kubenswrapper[4990]: E1203 13:25:54.249676 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a\": container with ID starting with 7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a not found: ID does not exist" containerID="7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.249712 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a"} err="failed to get container status \"7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a\": rpc error: code = NotFound desc = could not find container \"7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a\": container with ID starting with 7bab160ae0735f278ab4f51c438b82ea28b463007a012711ef0e0e79e6fc993a not found: ID does not exist" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.249737 4990 scope.go:117] "RemoveContainer" containerID="c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7" Dec 03 13:25:54 crc kubenswrapper[4990]: E1203 13:25:54.250035 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7\": container with ID starting with c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7 not found: ID does not exist" containerID="c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.250067 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7"} err="failed to get container status \"c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7\": rpc error: code = NotFound desc = could not find container \"c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7\": container with ID starting with c958fe7c313ac20a3fba5abb5157e2ea14d5d4ea36307e478abe42223be22fe7 not found: ID does not exist" Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.471760 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrnhf"] Dec 03 13:25:54 crc kubenswrapper[4990]: I1203 13:25:54.480985 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vrnhf"] Dec 03 13:25:56 crc kubenswrapper[4990]: I1203 13:25:56.286302 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" path="/var/lib/kubelet/pods/a8619cf9-e617-41e4-8c91-56fbd9c71f20/volumes" Dec 03 13:25:59 crc kubenswrapper[4990]: I1203 13:25:59.655740 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrb2f"] Dec 03 13:25:59 crc kubenswrapper[4990]: I1203 13:25:59.656364 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrb2f" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="registry-server" containerID="cri-o://28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62" gracePeriod=2 Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.137176 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.222203 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c2sd\" (UniqueName: \"kubernetes.io/projected/bb300e54-c925-4f74-97d5-8a189217b86b-kube-api-access-7c2sd\") pod \"bb300e54-c925-4f74-97d5-8a189217b86b\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.222432 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-utilities\") pod \"bb300e54-c925-4f74-97d5-8a189217b86b\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.222526 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-catalog-content\") pod \"bb300e54-c925-4f74-97d5-8a189217b86b\" (UID: \"bb300e54-c925-4f74-97d5-8a189217b86b\") " Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.226388 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-utilities" (OuterVolumeSpecName: "utilities") pod "bb300e54-c925-4f74-97d5-8a189217b86b" (UID: "bb300e54-c925-4f74-97d5-8a189217b86b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.270688 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb300e54-c925-4f74-97d5-8a189217b86b-kube-api-access-7c2sd" (OuterVolumeSpecName: "kube-api-access-7c2sd") pod "bb300e54-c925-4f74-97d5-8a189217b86b" (UID: "bb300e54-c925-4f74-97d5-8a189217b86b"). InnerVolumeSpecName "kube-api-access-7c2sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.273107 4990 generic.go:334] "Generic (PLEG): container finished" podID="bb300e54-c925-4f74-97d5-8a189217b86b" containerID="28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62" exitCode=0 Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.273186 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrb2f" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.304677 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerDied","Data":"28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62"} Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.304772 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrb2f" event={"ID":"bb300e54-c925-4f74-97d5-8a189217b86b","Type":"ContainerDied","Data":"6b66b882e7861334b56bac12735bb8074d61c9b806160387d1b87c93cfd94f67"} Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.304801 4990 scope.go:117] "RemoveContainer" containerID="28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.326032 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.326081 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c2sd\" (UniqueName: \"kubernetes.io/projected/bb300e54-c925-4f74-97d5-8a189217b86b-kube-api-access-7c2sd\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.345118 4990 scope.go:117] "RemoveContainer" containerID="83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.369622 4990 scope.go:117] "RemoveContainer" containerID="b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.410002 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb300e54-c925-4f74-97d5-8a189217b86b" (UID: "bb300e54-c925-4f74-97d5-8a189217b86b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.426797 4990 scope.go:117] "RemoveContainer" containerID="28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62" Dec 03 13:26:00 crc kubenswrapper[4990]: E1203 13:26:00.428207 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62\": container with ID starting with 28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62 not found: ID does not exist" containerID="28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.428251 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62"} err="failed to get container status \"28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62\": rpc error: code = NotFound desc = could not find container \"28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62\": container with ID starting with 28c1a94cf9476de5c893db30f9ab8b953852d63f722ddb7047cdaa837efd5f62 not found: ID does not exist" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.428280 4990 scope.go:117] "RemoveContainer" containerID="83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.428896 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb300e54-c925-4f74-97d5-8a189217b86b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:00 crc kubenswrapper[4990]: E1203 13:26:00.429252 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12\": container with ID starting with 83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12 not found: ID does not exist" containerID="83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.429294 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12"} err="failed to get container status \"83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12\": rpc error: code = NotFound desc = could not find container \"83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12\": container with ID starting with 83e35a0bc127e2ad2bde37f24f22beda86dfaa856e9896ef45baa86b238afe12 not found: ID does not exist" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.429327 4990 scope.go:117] "RemoveContainer" containerID="b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad" Dec 03 13:26:00 crc kubenswrapper[4990]: E1203 13:26:00.429641 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad\": container with ID starting with b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad not found: ID does not exist" containerID="b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.429678 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad"} err="failed to get container status \"b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad\": rpc error: code = NotFound desc = could not find container \"b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad\": container with ID starting with b538550b6e3dbbbd5b6a226fe6a48059f255b58d40ec3edd1ebd1656005b28ad not found: ID does not exist" Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.609653 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrb2f"] Dec 03 13:26:00 crc kubenswrapper[4990]: I1203 13:26:00.618547 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrb2f"] Dec 03 13:26:02 crc kubenswrapper[4990]: I1203 13:26:02.279065 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" path="/var/lib/kubelet/pods/bb300e54-c925-4f74-97d5-8a189217b86b/volumes" Dec 03 13:26:03 crc kubenswrapper[4990]: I1203 13:26:03.286069 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:26:03 crc kubenswrapper[4990]: I1203 13:26:03.286441 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.929387 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mtd68"] Dec 03 13:26:17 crc kubenswrapper[4990]: E1203 13:26:17.930575 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="extract-utilities" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.930593 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="extract-utilities" Dec 03 13:26:17 crc kubenswrapper[4990]: E1203 13:26:17.930616 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="extract-content" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.930624 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="extract-content" Dec 03 13:26:17 crc kubenswrapper[4990]: E1203 13:26:17.930639 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="extract-content" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.930647 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="extract-content" Dec 03 13:26:17 crc kubenswrapper[4990]: E1203 13:26:17.930663 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="extract-utilities" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.930671 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="extract-utilities" Dec 03 13:26:17 crc kubenswrapper[4990]: E1203 13:26:17.930695 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="registry-server" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.930702 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="registry-server" Dec 03 13:26:17 crc kubenswrapper[4990]: E1203 13:26:17.930726 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="registry-server" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.930733 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="registry-server" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.935268 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8619cf9-e617-41e4-8c91-56fbd9c71f20" containerName="registry-server" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.935336 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb300e54-c925-4f74-97d5-8a189217b86b" containerName="registry-server" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.937112 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.941318 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtd68"] Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.984797 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjf8d\" (UniqueName: \"kubernetes.io/projected/7443bbf2-61c9-4c42-b409-8858d02614e4-kube-api-access-xjf8d\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.984929 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-catalog-content\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:17 crc kubenswrapper[4990]: I1203 13:26:17.985335 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-utilities\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.086581 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjf8d\" (UniqueName: \"kubernetes.io/projected/7443bbf2-61c9-4c42-b409-8858d02614e4-kube-api-access-xjf8d\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.086644 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-catalog-content\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.086750 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-utilities\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.087307 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-catalog-content\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.087324 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-utilities\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.109988 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjf8d\" (UniqueName: \"kubernetes.io/projected/7443bbf2-61c9-4c42-b409-8858d02614e4-kube-api-access-xjf8d\") pod \"certified-operators-mtd68\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.259339 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:18 crc kubenswrapper[4990]: I1203 13:26:18.772338 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtd68"] Dec 03 13:26:19 crc kubenswrapper[4990]: I1203 13:26:19.478196 4990 generic.go:334] "Generic (PLEG): container finished" podID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerID="0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24" exitCode=0 Dec 03 13:26:19 crc kubenswrapper[4990]: I1203 13:26:19.478706 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerDied","Data":"0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24"} Dec 03 13:26:19 crc kubenswrapper[4990]: I1203 13:26:19.478764 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerStarted","Data":"2d30e6f77bc0aa5ca1cab98bb504eb5a2e52fc96cd6a690bd74477142972e8a0"} Dec 03 13:26:20 crc kubenswrapper[4990]: I1203 13:26:20.490469 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerStarted","Data":"d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14"} Dec 03 13:26:21 crc kubenswrapper[4990]: I1203 13:26:21.500994 4990 generic.go:334] "Generic (PLEG): container finished" podID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerID="d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14" exitCode=0 Dec 03 13:26:21 crc kubenswrapper[4990]: I1203 13:26:21.501048 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerDied","Data":"d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14"} Dec 03 13:26:22 crc kubenswrapper[4990]: I1203 13:26:22.514631 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerStarted","Data":"737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004"} Dec 03 13:26:22 crc kubenswrapper[4990]: I1203 13:26:22.540911 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mtd68" podStartSLOduration=3.076558209 podStartE2EDuration="5.540882276s" podCreationTimestamp="2025-12-03 13:26:17 +0000 UTC" firstStartedPulling="2025-12-03 13:26:19.481972535 +0000 UTC m=+2927.623883764" lastFinishedPulling="2025-12-03 13:26:21.946296582 +0000 UTC m=+2930.088207831" observedRunningTime="2025-12-03 13:26:22.530453169 +0000 UTC m=+2930.672364398" watchObservedRunningTime="2025-12-03 13:26:22.540882276 +0000 UTC m=+2930.682793515" Dec 03 13:26:28 crc kubenswrapper[4990]: I1203 13:26:28.259845 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:28 crc kubenswrapper[4990]: I1203 13:26:28.260501 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:28 crc kubenswrapper[4990]: I1203 13:26:28.316747 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:28 crc kubenswrapper[4990]: I1203 13:26:28.608063 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:28 crc kubenswrapper[4990]: I1203 13:26:28.654700 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtd68"] Dec 03 13:26:30 crc kubenswrapper[4990]: I1203 13:26:30.578257 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mtd68" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="registry-server" containerID="cri-o://737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004" gracePeriod=2 Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.078691 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.104384 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-utilities\") pod \"7443bbf2-61c9-4c42-b409-8858d02614e4\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.105770 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-catalog-content\") pod \"7443bbf2-61c9-4c42-b409-8858d02614e4\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.115871 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjf8d\" (UniqueName: \"kubernetes.io/projected/7443bbf2-61c9-4c42-b409-8858d02614e4-kube-api-access-xjf8d\") pod \"7443bbf2-61c9-4c42-b409-8858d02614e4\" (UID: \"7443bbf2-61c9-4c42-b409-8858d02614e4\") " Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.105701 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-utilities" (OuterVolumeSpecName: "utilities") pod "7443bbf2-61c9-4c42-b409-8858d02614e4" (UID: "7443bbf2-61c9-4c42-b409-8858d02614e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.116723 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.123288 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7443bbf2-61c9-4c42-b409-8858d02614e4-kube-api-access-xjf8d" (OuterVolumeSpecName: "kube-api-access-xjf8d") pod "7443bbf2-61c9-4c42-b409-8858d02614e4" (UID: "7443bbf2-61c9-4c42-b409-8858d02614e4"). InnerVolumeSpecName "kube-api-access-xjf8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.159974 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7443bbf2-61c9-4c42-b409-8858d02614e4" (UID: "7443bbf2-61c9-4c42-b409-8858d02614e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.218579 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7443bbf2-61c9-4c42-b409-8858d02614e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.218611 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjf8d\" (UniqueName: \"kubernetes.io/projected/7443bbf2-61c9-4c42-b409-8858d02614e4-kube-api-access-xjf8d\") on node \"crc\" DevicePath \"\"" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.588834 4990 generic.go:334] "Generic (PLEG): container finished" podID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerID="737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004" exitCode=0 Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.588866 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerDied","Data":"737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004"} Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.588935 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtd68" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.590263 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtd68" event={"ID":"7443bbf2-61c9-4c42-b409-8858d02614e4","Type":"ContainerDied","Data":"2d30e6f77bc0aa5ca1cab98bb504eb5a2e52fc96cd6a690bd74477142972e8a0"} Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.590279 4990 scope.go:117] "RemoveContainer" containerID="737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.618579 4990 scope.go:117] "RemoveContainer" containerID="d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.640534 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtd68"] Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.655205 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mtd68"] Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.662034 4990 scope.go:117] "RemoveContainer" containerID="0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.679152 4990 scope.go:117] "RemoveContainer" containerID="737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004" Dec 03 13:26:31 crc kubenswrapper[4990]: E1203 13:26:31.679625 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004\": container with ID starting with 737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004 not found: ID does not exist" containerID="737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.679655 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004"} err="failed to get container status \"737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004\": rpc error: code = NotFound desc = could not find container \"737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004\": container with ID starting with 737f42a6d8f15ae3754ea006d679f663377a108e6521b26570480b7ee1dc5004 not found: ID does not exist" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.679680 4990 scope.go:117] "RemoveContainer" containerID="d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14" Dec 03 13:26:31 crc kubenswrapper[4990]: E1203 13:26:31.680322 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14\": container with ID starting with d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14 not found: ID does not exist" containerID="d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.680342 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14"} err="failed to get container status \"d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14\": rpc error: code = NotFound desc = could not find container \"d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14\": container with ID starting with d2fd9b940a2e83ec5902028e2703a47547b983284641b581caa7a76c0a1add14 not found: ID does not exist" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.680377 4990 scope.go:117] "RemoveContainer" containerID="0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24" Dec 03 13:26:31 crc kubenswrapper[4990]: E1203 13:26:31.680608 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24\": container with ID starting with 0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24 not found: ID does not exist" containerID="0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24" Dec 03 13:26:31 crc kubenswrapper[4990]: I1203 13:26:31.680627 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24"} err="failed to get container status \"0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24\": rpc error: code = NotFound desc = could not find container \"0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24\": container with ID starting with 0f5fe1e21fc86a26ec970e10d71ee5cb3509b97efd0462011677ccda25ca7f24 not found: ID does not exist" Dec 03 13:26:32 crc kubenswrapper[4990]: I1203 13:26:32.276738 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" path="/var/lib/kubelet/pods/7443bbf2-61c9-4c42-b409-8858d02614e4/volumes" Dec 03 13:26:33 crc kubenswrapper[4990]: I1203 13:26:33.286716 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:26:33 crc kubenswrapper[4990]: I1203 13:26:33.286778 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:26:33 crc kubenswrapper[4990]: I1203 13:26:33.286826 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:26:33 crc kubenswrapper[4990]: I1203 13:26:33.287435 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:26:33 crc kubenswrapper[4990]: I1203 13:26:33.287533 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" gracePeriod=600 Dec 03 13:26:33 crc kubenswrapper[4990]: E1203 13:26:33.931724 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:26:34 crc kubenswrapper[4990]: I1203 13:26:34.622040 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" exitCode=0 Dec 03 13:26:34 crc kubenswrapper[4990]: I1203 13:26:34.622088 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5"} Dec 03 13:26:34 crc kubenswrapper[4990]: I1203 13:26:34.622130 4990 scope.go:117] "RemoveContainer" containerID="8e94b8c8d37ddcf489349bd8393b94b78192224584ef611211b491c2bcda4d09" Dec 03 13:26:34 crc kubenswrapper[4990]: I1203 13:26:34.622825 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:26:34 crc kubenswrapper[4990]: E1203 13:26:34.623124 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:26:47 crc kubenswrapper[4990]: I1203 13:26:47.264248 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:26:47 crc kubenswrapper[4990]: E1203 13:26:47.265242 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:27:01 crc kubenswrapper[4990]: I1203 13:27:01.264956 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:27:01 crc kubenswrapper[4990]: E1203 13:27:01.266244 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:27:13 crc kubenswrapper[4990]: I1203 13:27:13.264769 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:27:13 crc kubenswrapper[4990]: E1203 13:27:13.265897 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:27:27 crc kubenswrapper[4990]: I1203 13:27:27.265864 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:27:27 crc kubenswrapper[4990]: E1203 13:27:27.267206 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:27:42 crc kubenswrapper[4990]: I1203 13:27:42.283691 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:27:42 crc kubenswrapper[4990]: E1203 13:27:42.288158 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:27:57 crc kubenswrapper[4990]: I1203 13:27:57.265615 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:27:57 crc kubenswrapper[4990]: E1203 13:27:57.266714 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:28:06 crc kubenswrapper[4990]: I1203 13:28:06.720353 4990 generic.go:334] "Generic (PLEG): container finished" podID="4191ee3d-7722-4690-938c-7b8d8478589e" containerID="048dd83e63c4f1ac2c8da86ec3401d92a31a3c9d957d8a01536389b431972d24" exitCode=0 Dec 03 13:28:06 crc kubenswrapper[4990]: I1203 13:28:06.720477 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" event={"ID":"4191ee3d-7722-4690-938c-7b8d8478589e","Type":"ContainerDied","Data":"048dd83e63c4f1ac2c8da86ec3401d92a31a3c9d957d8a01536389b431972d24"} Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.132425 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192004 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b72v6\" (UniqueName: \"kubernetes.io/projected/4191ee3d-7722-4690-938c-7b8d8478589e-kube-api-access-b72v6\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192096 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-combined-ca-bundle\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192166 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-1\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192227 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-ssh-key\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192270 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-inventory\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192317 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-1\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192347 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-0\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192383 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4191ee3d-7722-4690-938c-7b8d8478589e-nova-extra-config-0\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.192427 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-0\") pod \"4191ee3d-7722-4690-938c-7b8d8478589e\" (UID: \"4191ee3d-7722-4690-938c-7b8d8478589e\") " Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.197990 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4191ee3d-7722-4690-938c-7b8d8478589e-kube-api-access-b72v6" (OuterVolumeSpecName: "kube-api-access-b72v6") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "kube-api-access-b72v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.198143 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.220571 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.221042 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.222246 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-inventory" (OuterVolumeSpecName: "inventory") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.227836 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4191ee3d-7722-4690-938c-7b8d8478589e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.236356 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.242425 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.252373 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4191ee3d-7722-4690-938c-7b8d8478589e" (UID: "4191ee3d-7722-4690-938c-7b8d8478589e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.266568 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:28:08 crc kubenswrapper[4990]: E1203 13:28:08.267002 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.295911 4990 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.296157 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.296248 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.296328 4990 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.296492 4990 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.297003 4990 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4191ee3d-7722-4690-938c-7b8d8478589e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.297908 4990 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.298854 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b72v6\" (UniqueName: \"kubernetes.io/projected/4191ee3d-7722-4690-938c-7b8d8478589e-kube-api-access-b72v6\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.298906 4990 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4191ee3d-7722-4690-938c-7b8d8478589e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.738416 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" event={"ID":"4191ee3d-7722-4690-938c-7b8d8478589e","Type":"ContainerDied","Data":"58a4fc7ce39faf10ef75ba642911d151c8c151722fb53fe487ee38fbe17bdb0e"} Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.738481 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58a4fc7ce39faf10ef75ba642911d151c8c151722fb53fe487ee38fbe17bdb0e" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.738543 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hhvts" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.866500 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9"] Dec 03 13:28:08 crc kubenswrapper[4990]: E1203 13:28:08.870138 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="extract-utilities" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.870193 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="extract-utilities" Dec 03 13:28:08 crc kubenswrapper[4990]: E1203 13:28:08.870259 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="registry-server" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.870269 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="registry-server" Dec 03 13:28:08 crc kubenswrapper[4990]: E1203 13:28:08.870307 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4191ee3d-7722-4690-938c-7b8d8478589e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.870316 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="4191ee3d-7722-4690-938c-7b8d8478589e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 13:28:08 crc kubenswrapper[4990]: E1203 13:28:08.870373 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="extract-content" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.870404 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="extract-content" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.877174 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="4191ee3d-7722-4690-938c-7b8d8478589e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.889439 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="7443bbf2-61c9-4c42-b409-8858d02614e4" containerName="registry-server" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.892958 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9"] Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.893367 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.900194 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j7p7b" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.900510 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.900552 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.900522 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 13:28:08 crc kubenswrapper[4990]: I1203 13:28:08.901627 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.012924 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.012993 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf7j7\" (UniqueName: \"kubernetes.io/projected/882b0d7e-4fd5-4462-ba56-091dbc993c0b-kube-api-access-xf7j7\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.013021 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.013069 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.013089 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.013133 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.013158 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.114628 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.114940 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.115108 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.115209 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.115481 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.115587 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf7j7\" (UniqueName: \"kubernetes.io/projected/882b0d7e-4fd5-4462-ba56-091dbc993c0b-kube-api-access-xf7j7\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.115689 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.120821 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.121098 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.121711 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.122048 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.122146 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.122858 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.141790 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf7j7\" (UniqueName: \"kubernetes.io/projected/882b0d7e-4fd5-4462-ba56-091dbc993c0b-kube-api-access-xf7j7\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.217863 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:28:09 crc kubenswrapper[4990]: I1203 13:28:09.778248 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9"] Dec 03 13:28:10 crc kubenswrapper[4990]: I1203 13:28:10.756674 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" event={"ID":"882b0d7e-4fd5-4462-ba56-091dbc993c0b","Type":"ContainerStarted","Data":"c13acc4e5691e743092a23fcc259a838d5e5c0a609b58f0f16fb20eba5641181"} Dec 03 13:28:11 crc kubenswrapper[4990]: I1203 13:28:11.771097 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" event={"ID":"882b0d7e-4fd5-4462-ba56-091dbc993c0b","Type":"ContainerStarted","Data":"57bf20ff0c7db260e187c33a80111a65625075304132ca8ff6a93d857674443e"} Dec 03 13:28:11 crc kubenswrapper[4990]: I1203 13:28:11.793118 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" podStartSLOduration=2.839822401 podStartE2EDuration="3.793099369s" podCreationTimestamp="2025-12-03 13:28:08 +0000 UTC" firstStartedPulling="2025-12-03 13:28:09.787860302 +0000 UTC m=+3037.929771531" lastFinishedPulling="2025-12-03 13:28:10.74113727 +0000 UTC m=+3038.883048499" observedRunningTime="2025-12-03 13:28:11.787111501 +0000 UTC m=+3039.929022740" watchObservedRunningTime="2025-12-03 13:28:11.793099369 +0000 UTC m=+3039.935010608" Dec 03 13:28:22 crc kubenswrapper[4990]: I1203 13:28:22.270688 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:28:22 crc kubenswrapper[4990]: E1203 13:28:22.271925 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:28:35 crc kubenswrapper[4990]: I1203 13:28:35.265354 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:28:35 crc kubenswrapper[4990]: E1203 13:28:35.266142 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:28:49 crc kubenswrapper[4990]: I1203 13:28:49.264509 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:28:49 crc kubenswrapper[4990]: E1203 13:28:49.265366 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:29:02 crc kubenswrapper[4990]: I1203 13:29:02.274635 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:29:02 crc kubenswrapper[4990]: E1203 13:29:02.275577 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:29:16 crc kubenswrapper[4990]: I1203 13:29:16.267482 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:29:16 crc kubenswrapper[4990]: E1203 13:29:16.268696 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:29:30 crc kubenswrapper[4990]: I1203 13:29:30.264710 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:29:30 crc kubenswrapper[4990]: E1203 13:29:30.265714 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:29:44 crc kubenswrapper[4990]: I1203 13:29:44.264522 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:29:44 crc kubenswrapper[4990]: E1203 13:29:44.269343 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:29:55 crc kubenswrapper[4990]: I1203 13:29:55.268581 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:29:55 crc kubenswrapper[4990]: E1203 13:29:55.271969 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.156902 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2"] Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.159056 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.161005 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.165393 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.182145 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2"] Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.249065 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-config-volume\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.249127 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw5xl\" (UniqueName: \"kubernetes.io/projected/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-kube-api-access-lw5xl\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.249230 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-secret-volume\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.351327 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-config-volume\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.351400 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw5xl\" (UniqueName: \"kubernetes.io/projected/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-kube-api-access-lw5xl\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.351447 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-secret-volume\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.352304 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-config-volume\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.357138 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-secret-volume\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.368025 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw5xl\" (UniqueName: \"kubernetes.io/projected/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-kube-api-access-lw5xl\") pod \"collect-profiles-29412810-7bzv2\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.481427 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:00 crc kubenswrapper[4990]: I1203 13:30:00.943981 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2"] Dec 03 13:30:01 crc kubenswrapper[4990]: I1203 13:30:01.822939 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" event={"ID":"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1","Type":"ContainerStarted","Data":"430ae9fcf3fb8cd4ae08ee1846718e7a9618ba06804406d16321d962094bd75b"} Dec 03 13:30:02 crc kubenswrapper[4990]: E1203 13:30:02.266612 4990 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05ea3f9a_e6b2_4242_b2e7_e283ed53cbe1.slice/crio-a1ec6ecb3c5f57d8de849b20400c53457663648f2423daa6a0e4046e5590f2dc.scope\": RecentStats: unable to find data in memory cache]" Dec 03 13:30:02 crc kubenswrapper[4990]: I1203 13:30:02.834777 4990 generic.go:334] "Generic (PLEG): container finished" podID="05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" containerID="a1ec6ecb3c5f57d8de849b20400c53457663648f2423daa6a0e4046e5590f2dc" exitCode=0 Dec 03 13:30:02 crc kubenswrapper[4990]: I1203 13:30:02.834897 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" event={"ID":"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1","Type":"ContainerDied","Data":"a1ec6ecb3c5f57d8de849b20400c53457663648f2423daa6a0e4046e5590f2dc"} Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.172715 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.237880 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-secret-volume\") pod \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.238041 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-config-volume\") pod \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.238106 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw5xl\" (UniqueName: \"kubernetes.io/projected/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-kube-api-access-lw5xl\") pod \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\" (UID: \"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1\") " Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.238884 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-config-volume" (OuterVolumeSpecName: "config-volume") pod "05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" (UID: "05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.282324 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" (UID: "05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.283269 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-kube-api-access-lw5xl" (OuterVolumeSpecName: "kube-api-access-lw5xl") pod "05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" (UID: "05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1"). InnerVolumeSpecName "kube-api-access-lw5xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.364175 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.364241 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.364265 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw5xl\" (UniqueName: \"kubernetes.io/projected/05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1-kube-api-access-lw5xl\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.937943 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" event={"ID":"05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1","Type":"ContainerDied","Data":"430ae9fcf3fb8cd4ae08ee1846718e7a9618ba06804406d16321d962094bd75b"} Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.938352 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="430ae9fcf3fb8cd4ae08ee1846718e7a9618ba06804406d16321d962094bd75b" Dec 03 13:30:04 crc kubenswrapper[4990]: I1203 13:30:04.938124 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412810-7bzv2" Dec 03 13:30:05 crc kubenswrapper[4990]: I1203 13:30:05.272558 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h"] Dec 03 13:30:05 crc kubenswrapper[4990]: I1203 13:30:05.283092 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412765-jzt4h"] Dec 03 13:30:06 crc kubenswrapper[4990]: I1203 13:30:06.275336 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="471bdda0-0146-48d8-af64-0cfd178fdd48" path="/var/lib/kubelet/pods/471bdda0-0146-48d8-af64-0cfd178fdd48/volumes" Dec 03 13:30:10 crc kubenswrapper[4990]: I1203 13:30:10.265907 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:30:10 crc kubenswrapper[4990]: E1203 13:30:10.267354 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:30:22 crc kubenswrapper[4990]: I1203 13:30:22.274299 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:30:22 crc kubenswrapper[4990]: E1203 13:30:22.275399 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:30:26 crc kubenswrapper[4990]: I1203 13:30:26.551062 4990 scope.go:117] "RemoveContainer" containerID="7f48c4ef412c45a8d91880a69311c45ef0381639de6bad984188b4f929a418ba" Dec 03 13:30:28 crc kubenswrapper[4990]: I1203 13:30:28.175090 4990 generic.go:334] "Generic (PLEG): container finished" podID="882b0d7e-4fd5-4462-ba56-091dbc993c0b" containerID="57bf20ff0c7db260e187c33a80111a65625075304132ca8ff6a93d857674443e" exitCode=0 Dec 03 13:30:28 crc kubenswrapper[4990]: I1203 13:30:28.175144 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" event={"ID":"882b0d7e-4fd5-4462-ba56-091dbc993c0b","Type":"ContainerDied","Data":"57bf20ff0c7db260e187c33a80111a65625075304132ca8ff6a93d857674443e"} Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.606949 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.792910 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-0\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.793013 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-telemetry-combined-ca-bundle\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.793037 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-2\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.793070 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf7j7\" (UniqueName: \"kubernetes.io/projected/882b0d7e-4fd5-4462-ba56-091dbc993c0b-kube-api-access-xf7j7\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.793193 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-inventory\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.793236 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-1\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.793266 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ssh-key\") pod \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\" (UID: \"882b0d7e-4fd5-4462-ba56-091dbc993c0b\") " Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.809518 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.811633 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/882b0d7e-4fd5-4462-ba56-091dbc993c0b-kube-api-access-xf7j7" (OuterVolumeSpecName: "kube-api-access-xf7j7") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "kube-api-access-xf7j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.828765 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.839836 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.841907 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-inventory" (OuterVolumeSpecName: "inventory") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.849212 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.856754 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "882b0d7e-4fd5-4462-ba56-091dbc993c0b" (UID: "882b0d7e-4fd5-4462-ba56-091dbc993c0b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897725 4990 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897770 4990 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897782 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897792 4990 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897803 4990 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897813 4990 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/882b0d7e-4fd5-4462-ba56-091dbc993c0b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:29 crc kubenswrapper[4990]: I1203 13:30:29.897824 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf7j7\" (UniqueName: \"kubernetes.io/projected/882b0d7e-4fd5-4462-ba56-091dbc993c0b-kube-api-access-xf7j7\") on node \"crc\" DevicePath \"\"" Dec 03 13:30:30 crc kubenswrapper[4990]: I1203 13:30:30.201025 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" event={"ID":"882b0d7e-4fd5-4462-ba56-091dbc993c0b","Type":"ContainerDied","Data":"c13acc4e5691e743092a23fcc259a838d5e5c0a609b58f0f16fb20eba5641181"} Dec 03 13:30:30 crc kubenswrapper[4990]: I1203 13:30:30.201470 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c13acc4e5691e743092a23fcc259a838d5e5c0a609b58f0f16fb20eba5641181" Dec 03 13:30:30 crc kubenswrapper[4990]: I1203 13:30:30.201093 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9" Dec 03 13:30:37 crc kubenswrapper[4990]: I1203 13:30:37.265141 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:30:37 crc kubenswrapper[4990]: E1203 13:30:37.265999 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:30:52 crc kubenswrapper[4990]: I1203 13:30:52.271993 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:30:52 crc kubenswrapper[4990]: E1203 13:30:52.272924 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:31:05 crc kubenswrapper[4990]: I1203 13:31:05.264642 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:31:05 crc kubenswrapper[4990]: E1203 13:31:05.265324 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:31:20 crc kubenswrapper[4990]: I1203 13:31:20.265072 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:31:20 crc kubenswrapper[4990]: E1203 13:31:20.265932 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.495707 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 13:31:27 crc kubenswrapper[4990]: E1203 13:31:27.496801 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" containerName="collect-profiles" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.496818 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" containerName="collect-profiles" Dec 03 13:31:27 crc kubenswrapper[4990]: E1203 13:31:27.496839 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882b0d7e-4fd5-4462-ba56-091dbc993c0b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.496846 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="882b0d7e-4fd5-4462-ba56-091dbc993c0b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.497038 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ea3f9a-e6b2-4242-b2e7-e283ed53cbe1" containerName="collect-profiles" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.497054 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="882b0d7e-4fd5-4462-ba56-091dbc993c0b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.499103 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.502364 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.502554 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.502607 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.502726 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-c2cfh" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.516369 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.648544 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.648604 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.648654 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c68bt\" (UniqueName: \"kubernetes.io/projected/f375426e-6095-4621-ab29-6c7324da86e7-kube-api-access-c68bt\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.648932 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.649095 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.649197 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.649317 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.649553 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.649600 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-config-data\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.752538 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.752691 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.752796 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.752983 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.753204 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.753373 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.753792 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.753903 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-config-data\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.754828 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.755022 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.755078 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c68bt\" (UniqueName: \"kubernetes.io/projected/f375426e-6095-4621-ab29-6c7324da86e7-kube-api-access-c68bt\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.755413 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.756073 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-config-data\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.756437 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.760896 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.762627 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.763590 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.773029 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c68bt\" (UniqueName: \"kubernetes.io/projected/f375426e-6095-4621-ab29-6c7324da86e7-kube-api-access-c68bt\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.789754 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " pod="openstack/tempest-tests-tempest" Dec 03 13:31:27 crc kubenswrapper[4990]: I1203 13:31:27.826874 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 13:31:28 crc kubenswrapper[4990]: I1203 13:31:28.301216 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 13:31:28 crc kubenswrapper[4990]: I1203 13:31:28.308280 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:31:28 crc kubenswrapper[4990]: I1203 13:31:28.529858 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f375426e-6095-4621-ab29-6c7324da86e7","Type":"ContainerStarted","Data":"ce20b67ff8e6e9c2a5f66d2043554b636d5074adafd43a65ce7891b54fdc56c9"} Dec 03 13:31:31 crc kubenswrapper[4990]: I1203 13:31:31.265242 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:31:31 crc kubenswrapper[4990]: E1203 13:31:31.265888 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:31:45 crc kubenswrapper[4990]: I1203 13:31:45.265620 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.484880 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hw8lc"] Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.489374 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.513987 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw8lc"] Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.530426 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-catalog-content\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.530527 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2rmh\" (UniqueName: \"kubernetes.io/projected/c662e65c-68b9-4a59-b925-f650ab846bd1-kube-api-access-c2rmh\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.530667 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-utilities\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.632931 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-utilities\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.632994 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-catalog-content\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.633014 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2rmh\" (UniqueName: \"kubernetes.io/projected/c662e65c-68b9-4a59-b925-f650ab846bd1-kube-api-access-c2rmh\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.633493 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-catalog-content\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.636347 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-utilities\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.653952 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2rmh\" (UniqueName: \"kubernetes.io/projected/c662e65c-68b9-4a59-b925-f650ab846bd1-kube-api-access-c2rmh\") pod \"redhat-marketplace-hw8lc\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:04 crc kubenswrapper[4990]: I1203 13:32:04.825344 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:09 crc kubenswrapper[4990]: E1203 13:32:09.995737 4990 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 13:32:10 crc kubenswrapper[4990]: E1203 13:32:10.000348 4990 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c68bt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(f375426e-6095-4621-ab29-6c7324da86e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 13:32:10 crc kubenswrapper[4990]: E1203 13:32:10.001627 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="f375426e-6095-4621-ab29-6c7324da86e7" Dec 03 13:32:10 crc kubenswrapper[4990]: I1203 13:32:10.375267 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw8lc"] Dec 03 13:32:10 crc kubenswrapper[4990]: I1203 13:32:10.981773 4990 generic.go:334] "Generic (PLEG): container finished" podID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerID="a78146852aef772b16ff6b1b1193bd358c00f410f45fb9d98975a8ab4ed79ec8" exitCode=0 Dec 03 13:32:10 crc kubenswrapper[4990]: I1203 13:32:10.981848 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw8lc" event={"ID":"c662e65c-68b9-4a59-b925-f650ab846bd1","Type":"ContainerDied","Data":"a78146852aef772b16ff6b1b1193bd358c00f410f45fb9d98975a8ab4ed79ec8"} Dec 03 13:32:10 crc kubenswrapper[4990]: I1203 13:32:10.982288 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw8lc" event={"ID":"c662e65c-68b9-4a59-b925-f650ab846bd1","Type":"ContainerStarted","Data":"175b694db7d3ca4f366408ab0bfa3eb06e2dda741f13687e488066f9c5025865"} Dec 03 13:32:10 crc kubenswrapper[4990]: I1203 13:32:10.994686 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"759279b6d38b81a7f11d8d2f79cf938abb09838e4952a8a80b89c5ee5b26fce1"} Dec 03 13:32:11 crc kubenswrapper[4990]: E1203 13:32:11.003641 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="f375426e-6095-4621-ab29-6c7324da86e7" Dec 03 13:32:13 crc kubenswrapper[4990]: I1203 13:32:13.016852 4990 generic.go:334] "Generic (PLEG): container finished" podID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerID="71aa83832dc8fa95f41e78ecc034310f7227aa608c508209d06a8c82fb68c244" exitCode=0 Dec 03 13:32:13 crc kubenswrapper[4990]: I1203 13:32:13.016956 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw8lc" event={"ID":"c662e65c-68b9-4a59-b925-f650ab846bd1","Type":"ContainerDied","Data":"71aa83832dc8fa95f41e78ecc034310f7227aa608c508209d06a8c82fb68c244"} Dec 03 13:32:14 crc kubenswrapper[4990]: I1203 13:32:14.029213 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw8lc" event={"ID":"c662e65c-68b9-4a59-b925-f650ab846bd1","Type":"ContainerStarted","Data":"d91d37582f7e94a75feaf54e8b31b09de9c6d173bc18c8ec9248a99f58525030"} Dec 03 13:32:14 crc kubenswrapper[4990]: I1203 13:32:14.050816 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hw8lc" podStartSLOduration=7.532352732 podStartE2EDuration="10.050795297s" podCreationTimestamp="2025-12-03 13:32:04 +0000 UTC" firstStartedPulling="2025-12-03 13:32:10.989869802 +0000 UTC m=+3279.131781031" lastFinishedPulling="2025-12-03 13:32:13.508312337 +0000 UTC m=+3281.650223596" observedRunningTime="2025-12-03 13:32:14.046612905 +0000 UTC m=+3282.188524154" watchObservedRunningTime="2025-12-03 13:32:14.050795297 +0000 UTC m=+3282.192706536" Dec 03 13:32:14 crc kubenswrapper[4990]: I1203 13:32:14.826188 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:14 crc kubenswrapper[4990]: I1203 13:32:14.826270 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:15 crc kubenswrapper[4990]: I1203 13:32:15.883868 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-hw8lc" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="registry-server" probeResult="failure" output=< Dec 03 13:32:15 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 13:32:15 crc kubenswrapper[4990]: > Dec 03 13:32:22 crc kubenswrapper[4990]: I1203 13:32:22.890095 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 13:32:24 crc kubenswrapper[4990]: I1203 13:32:24.125428 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f375426e-6095-4621-ab29-6c7324da86e7","Type":"ContainerStarted","Data":"630e02c903627ad97e6cc587cad2680a000e819c3908ebda0e544b5ef2fcf983"} Dec 03 13:32:24 crc kubenswrapper[4990]: I1203 13:32:24.152118 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.576183388 podStartE2EDuration="58.152098902s" podCreationTimestamp="2025-12-03 13:31:26 +0000 UTC" firstStartedPulling="2025-12-03 13:31:28.308000691 +0000 UTC m=+3236.449911920" lastFinishedPulling="2025-12-03 13:32:22.883916205 +0000 UTC m=+3291.025827434" observedRunningTime="2025-12-03 13:32:24.146337839 +0000 UTC m=+3292.288249068" watchObservedRunningTime="2025-12-03 13:32:24.152098902 +0000 UTC m=+3292.294010131" Dec 03 13:32:24 crc kubenswrapper[4990]: I1203 13:32:24.904653 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:24 crc kubenswrapper[4990]: I1203 13:32:24.972354 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:25 crc kubenswrapper[4990]: I1203 13:32:25.141613 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw8lc"] Dec 03 13:32:26 crc kubenswrapper[4990]: I1203 13:32:26.144400 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hw8lc" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="registry-server" containerID="cri-o://d91d37582f7e94a75feaf54e8b31b09de9c6d173bc18c8ec9248a99f58525030" gracePeriod=2 Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.209144 4990 generic.go:334] "Generic (PLEG): container finished" podID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerID="d91d37582f7e94a75feaf54e8b31b09de9c6d173bc18c8ec9248a99f58525030" exitCode=0 Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.209508 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw8lc" event={"ID":"c662e65c-68b9-4a59-b925-f650ab846bd1","Type":"ContainerDied","Data":"d91d37582f7e94a75feaf54e8b31b09de9c6d173bc18c8ec9248a99f58525030"} Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.494661 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.612654 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-utilities\") pod \"c662e65c-68b9-4a59-b925-f650ab846bd1\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.612714 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-catalog-content\") pod \"c662e65c-68b9-4a59-b925-f650ab846bd1\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.612814 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2rmh\" (UniqueName: \"kubernetes.io/projected/c662e65c-68b9-4a59-b925-f650ab846bd1-kube-api-access-c2rmh\") pod \"c662e65c-68b9-4a59-b925-f650ab846bd1\" (UID: \"c662e65c-68b9-4a59-b925-f650ab846bd1\") " Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.613805 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-utilities" (OuterVolumeSpecName: "utilities") pod "c662e65c-68b9-4a59-b925-f650ab846bd1" (UID: "c662e65c-68b9-4a59-b925-f650ab846bd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.628115 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c662e65c-68b9-4a59-b925-f650ab846bd1-kube-api-access-c2rmh" (OuterVolumeSpecName: "kube-api-access-c2rmh") pod "c662e65c-68b9-4a59-b925-f650ab846bd1" (UID: "c662e65c-68b9-4a59-b925-f650ab846bd1"). InnerVolumeSpecName "kube-api-access-c2rmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.631556 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c662e65c-68b9-4a59-b925-f650ab846bd1" (UID: "c662e65c-68b9-4a59-b925-f650ab846bd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.714871 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.714927 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2rmh\" (UniqueName: \"kubernetes.io/projected/c662e65c-68b9-4a59-b925-f650ab846bd1-kube-api-access-c2rmh\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:27 crc kubenswrapper[4990]: I1203 13:32:27.714950 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c662e65c-68b9-4a59-b925-f650ab846bd1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.220086 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw8lc" event={"ID":"c662e65c-68b9-4a59-b925-f650ab846bd1","Type":"ContainerDied","Data":"175b694db7d3ca4f366408ab0bfa3eb06e2dda741f13687e488066f9c5025865"} Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.220421 4990 scope.go:117] "RemoveContainer" containerID="d91d37582f7e94a75feaf54e8b31b09de9c6d173bc18c8ec9248a99f58525030" Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.220211 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw8lc" Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.248937 4990 scope.go:117] "RemoveContainer" containerID="71aa83832dc8fa95f41e78ecc034310f7227aa608c508209d06a8c82fb68c244" Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.259548 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw8lc"] Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.276977 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw8lc"] Dec 03 13:32:28 crc kubenswrapper[4990]: I1203 13:32:28.287719 4990 scope.go:117] "RemoveContainer" containerID="a78146852aef772b16ff6b1b1193bd358c00f410f45fb9d98975a8ab4ed79ec8" Dec 03 13:32:30 crc kubenswrapper[4990]: I1203 13:32:30.276808 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" path="/var/lib/kubelet/pods/c662e65c-68b9-4a59-b925-f650ab846bd1/volumes" Dec 03 13:34:33 crc kubenswrapper[4990]: I1203 13:34:33.286608 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:34:33 crc kubenswrapper[4990]: I1203 13:34:33.287385 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:35:03 crc kubenswrapper[4990]: I1203 13:35:03.286546 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:35:03 crc kubenswrapper[4990]: I1203 13:35:03.287021 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.286870 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.287368 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.287413 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.288192 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"759279b6d38b81a7f11d8d2f79cf938abb09838e4952a8a80b89c5ee5b26fce1"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.288242 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://759279b6d38b81a7f11d8d2f79cf938abb09838e4952a8a80b89c5ee5b26fce1" gracePeriod=600 Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.963653 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="759279b6d38b81a7f11d8d2f79cf938abb09838e4952a8a80b89c5ee5b26fce1" exitCode=0 Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.963718 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"759279b6d38b81a7f11d8d2f79cf938abb09838e4952a8a80b89c5ee5b26fce1"} Dec 03 13:35:33 crc kubenswrapper[4990]: I1203 13:35:33.963844 4990 scope.go:117] "RemoveContainer" containerID="427c282e78737990b435cf5ddbf31607fcc1a0df596c93cc4f82a3f399b15bd5" Dec 03 13:35:34 crc kubenswrapper[4990]: I1203 13:35:34.974234 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173"} Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.500000 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-25k6b"] Dec 03 13:35:59 crc kubenswrapper[4990]: E1203 13:35:59.501122 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="registry-server" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.501141 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="registry-server" Dec 03 13:35:59 crc kubenswrapper[4990]: E1203 13:35:59.501172 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="extract-utilities" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.501180 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="extract-utilities" Dec 03 13:35:59 crc kubenswrapper[4990]: E1203 13:35:59.501200 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="extract-content" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.501208 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="extract-content" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.501499 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c662e65c-68b9-4a59-b925-f650ab846bd1" containerName="registry-server" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.503284 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.514638 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25k6b"] Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.687038 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-utilities\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.687087 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-catalog-content\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.687123 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m2wt\" (UniqueName: \"kubernetes.io/projected/e930d8f5-e952-49d7-bcbe-02a39ee37267-kube-api-access-7m2wt\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.789068 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-utilities\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.789410 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-catalog-content\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.789467 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m2wt\" (UniqueName: \"kubernetes.io/projected/e930d8f5-e952-49d7-bcbe-02a39ee37267-kube-api-access-7m2wt\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.789784 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-utilities\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.789880 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-catalog-content\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.815432 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m2wt\" (UniqueName: \"kubernetes.io/projected/e930d8f5-e952-49d7-bcbe-02a39ee37267-kube-api-access-7m2wt\") pod \"redhat-operators-25k6b\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:35:59 crc kubenswrapper[4990]: I1203 13:35:59.838559 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:00 crc kubenswrapper[4990]: I1203 13:36:00.338629 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25k6b"] Dec 03 13:36:01 crc kubenswrapper[4990]: I1203 13:36:01.271990 4990 generic.go:334] "Generic (PLEG): container finished" podID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerID="a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233" exitCode=0 Dec 03 13:36:01 crc kubenswrapper[4990]: I1203 13:36:01.272122 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerDied","Data":"a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233"} Dec 03 13:36:01 crc kubenswrapper[4990]: I1203 13:36:01.272234 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerStarted","Data":"fdf0a5320a3403a9ffc083a0c39d0fe9db128c4871b3b93582fe0c78fca76123"} Dec 03 13:36:03 crc kubenswrapper[4990]: I1203 13:36:03.291566 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerStarted","Data":"dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7"} Dec 03 13:36:04 crc kubenswrapper[4990]: I1203 13:36:04.301965 4990 generic.go:334] "Generic (PLEG): container finished" podID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerID="dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7" exitCode=0 Dec 03 13:36:04 crc kubenswrapper[4990]: I1203 13:36:04.302035 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerDied","Data":"dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7"} Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.271877 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f248l"] Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.279847 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.284199 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f248l"] Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.406704 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-catalog-content\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.406790 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-utilities\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.406940 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcwsm\" (UniqueName: \"kubernetes.io/projected/e81ddbb3-463e-4271-b0d6-baf34d74cb26-kube-api-access-rcwsm\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.509256 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-utilities\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.509407 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcwsm\" (UniqueName: \"kubernetes.io/projected/e81ddbb3-463e-4271-b0d6-baf34d74cb26-kube-api-access-rcwsm\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.509577 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-catalog-content\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.509856 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-utilities\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.509999 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-catalog-content\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.529639 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcwsm\" (UniqueName: \"kubernetes.io/projected/e81ddbb3-463e-4271-b0d6-baf34d74cb26-kube-api-access-rcwsm\") pod \"community-operators-f248l\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:05 crc kubenswrapper[4990]: I1203 13:36:05.601007 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:06 crc kubenswrapper[4990]: I1203 13:36:06.183218 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f248l"] Dec 03 13:36:06 crc kubenswrapper[4990]: I1203 13:36:06.319392 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f248l" event={"ID":"e81ddbb3-463e-4271-b0d6-baf34d74cb26","Type":"ContainerStarted","Data":"715f51e04296bec766ea5c631494fc5d4fbe29ccccff0d5c77ae856305e5b4f7"} Dec 03 13:36:08 crc kubenswrapper[4990]: I1203 13:36:08.349181 4990 generic.go:334] "Generic (PLEG): container finished" podID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerID="1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0" exitCode=0 Dec 03 13:36:08 crc kubenswrapper[4990]: I1203 13:36:08.349717 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f248l" event={"ID":"e81ddbb3-463e-4271-b0d6-baf34d74cb26","Type":"ContainerDied","Data":"1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0"} Dec 03 13:36:09 crc kubenswrapper[4990]: I1203 13:36:09.360586 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerStarted","Data":"6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555"} Dec 03 13:36:09 crc kubenswrapper[4990]: I1203 13:36:09.408977 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-25k6b" podStartSLOduration=3.111576751 podStartE2EDuration="10.408950892s" podCreationTimestamp="2025-12-03 13:35:59 +0000 UTC" firstStartedPulling="2025-12-03 13:36:01.273854271 +0000 UTC m=+3509.415765490" lastFinishedPulling="2025-12-03 13:36:08.571228412 +0000 UTC m=+3516.713139631" observedRunningTime="2025-12-03 13:36:09.405108928 +0000 UTC m=+3517.547020157" watchObservedRunningTime="2025-12-03 13:36:09.408950892 +0000 UTC m=+3517.550862121" Dec 03 13:36:09 crc kubenswrapper[4990]: I1203 13:36:09.839157 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:09 crc kubenswrapper[4990]: I1203 13:36:09.839488 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:10 crc kubenswrapper[4990]: I1203 13:36:10.895446 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-25k6b" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="registry-server" probeResult="failure" output=< Dec 03 13:36:10 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 13:36:10 crc kubenswrapper[4990]: > Dec 03 13:36:11 crc kubenswrapper[4990]: I1203 13:36:11.380948 4990 generic.go:334] "Generic (PLEG): container finished" podID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerID="36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7" exitCode=0 Dec 03 13:36:11 crc kubenswrapper[4990]: I1203 13:36:11.381016 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f248l" event={"ID":"e81ddbb3-463e-4271-b0d6-baf34d74cb26","Type":"ContainerDied","Data":"36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7"} Dec 03 13:36:13 crc kubenswrapper[4990]: I1203 13:36:13.400237 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f248l" event={"ID":"e81ddbb3-463e-4271-b0d6-baf34d74cb26","Type":"ContainerStarted","Data":"69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9"} Dec 03 13:36:13 crc kubenswrapper[4990]: I1203 13:36:13.423427 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f248l" podStartSLOduration=5.323736377 podStartE2EDuration="8.423406891s" podCreationTimestamp="2025-12-03 13:36:05 +0000 UTC" firstStartedPulling="2025-12-03 13:36:09.362342051 +0000 UTC m=+3517.504253280" lastFinishedPulling="2025-12-03 13:36:12.462012565 +0000 UTC m=+3520.603923794" observedRunningTime="2025-12-03 13:36:13.416533665 +0000 UTC m=+3521.558444894" watchObservedRunningTime="2025-12-03 13:36:13.423406891 +0000 UTC m=+3521.565318120" Dec 03 13:36:15 crc kubenswrapper[4990]: I1203 13:36:15.601859 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:15 crc kubenswrapper[4990]: I1203 13:36:15.602158 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:15 crc kubenswrapper[4990]: I1203 13:36:15.684600 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:19 crc kubenswrapper[4990]: I1203 13:36:19.941066 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:19 crc kubenswrapper[4990]: I1203 13:36:19.995559 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:20 crc kubenswrapper[4990]: I1203 13:36:20.208197 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25k6b"] Dec 03 13:36:21 crc kubenswrapper[4990]: I1203 13:36:21.465604 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-25k6b" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="registry-server" containerID="cri-o://6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555" gracePeriod=2 Dec 03 13:36:21 crc kubenswrapper[4990]: I1203 13:36:21.942464 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.058754 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-utilities\") pod \"e930d8f5-e952-49d7-bcbe-02a39ee37267\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.058830 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m2wt\" (UniqueName: \"kubernetes.io/projected/e930d8f5-e952-49d7-bcbe-02a39ee37267-kube-api-access-7m2wt\") pod \"e930d8f5-e952-49d7-bcbe-02a39ee37267\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.059121 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-catalog-content\") pod \"e930d8f5-e952-49d7-bcbe-02a39ee37267\" (UID: \"e930d8f5-e952-49d7-bcbe-02a39ee37267\") " Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.059628 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-utilities" (OuterVolumeSpecName: "utilities") pod "e930d8f5-e952-49d7-bcbe-02a39ee37267" (UID: "e930d8f5-e952-49d7-bcbe-02a39ee37267"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.077061 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e930d8f5-e952-49d7-bcbe-02a39ee37267-kube-api-access-7m2wt" (OuterVolumeSpecName: "kube-api-access-7m2wt") pod "e930d8f5-e952-49d7-bcbe-02a39ee37267" (UID: "e930d8f5-e952-49d7-bcbe-02a39ee37267"). InnerVolumeSpecName "kube-api-access-7m2wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.161339 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.161377 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m2wt\" (UniqueName: \"kubernetes.io/projected/e930d8f5-e952-49d7-bcbe-02a39ee37267-kube-api-access-7m2wt\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.176110 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e930d8f5-e952-49d7-bcbe-02a39ee37267" (UID: "e930d8f5-e952-49d7-bcbe-02a39ee37267"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.263568 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e930d8f5-e952-49d7-bcbe-02a39ee37267-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.476128 4990 generic.go:334] "Generic (PLEG): container finished" podID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerID="6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555" exitCode=0 Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.476176 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerDied","Data":"6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555"} Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.476194 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25k6b" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.476227 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25k6b" event={"ID":"e930d8f5-e952-49d7-bcbe-02a39ee37267","Type":"ContainerDied","Data":"fdf0a5320a3403a9ffc083a0c39d0fe9db128c4871b3b93582fe0c78fca76123"} Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.476253 4990 scope.go:117] "RemoveContainer" containerID="6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.500721 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25k6b"] Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.505137 4990 scope.go:117] "RemoveContainer" containerID="dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.512294 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-25k6b"] Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.531797 4990 scope.go:117] "RemoveContainer" containerID="a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.572144 4990 scope.go:117] "RemoveContainer" containerID="6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555" Dec 03 13:36:22 crc kubenswrapper[4990]: E1203 13:36:22.572939 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555\": container with ID starting with 6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555 not found: ID does not exist" containerID="6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.572979 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555"} err="failed to get container status \"6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555\": rpc error: code = NotFound desc = could not find container \"6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555\": container with ID starting with 6eb076740e6dfa81e9ac87248cf2aec1a61fc787852f8ca20f41df7316a0b555 not found: ID does not exist" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.573005 4990 scope.go:117] "RemoveContainer" containerID="dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7" Dec 03 13:36:22 crc kubenswrapper[4990]: E1203 13:36:22.573423 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7\": container with ID starting with dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7 not found: ID does not exist" containerID="dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.573481 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7"} err="failed to get container status \"dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7\": rpc error: code = NotFound desc = could not find container \"dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7\": container with ID starting with dce8dbdf92337cdf1dc287c87ac2ced9c1b091ee769f0070ebbee47f66b41ec7 not found: ID does not exist" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.573510 4990 scope.go:117] "RemoveContainer" containerID="a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233" Dec 03 13:36:22 crc kubenswrapper[4990]: E1203 13:36:22.573844 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233\": container with ID starting with a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233 not found: ID does not exist" containerID="a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233" Dec 03 13:36:22 crc kubenswrapper[4990]: I1203 13:36:22.573871 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233"} err="failed to get container status \"a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233\": rpc error: code = NotFound desc = could not find container \"a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233\": container with ID starting with a4b6491e7ff569c52943ea2f6ae63c7a090b86721e2b1dffa071ac5011867233 not found: ID does not exist" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.276100 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" path="/var/lib/kubelet/pods/e930d8f5-e952-49d7-bcbe-02a39ee37267/volumes" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.649444 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k7rt9"] Dec 03 13:36:24 crc kubenswrapper[4990]: E1203 13:36:24.650837 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="registry-server" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.650941 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="registry-server" Dec 03 13:36:24 crc kubenswrapper[4990]: E1203 13:36:24.651074 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="extract-utilities" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.651162 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="extract-utilities" Dec 03 13:36:24 crc kubenswrapper[4990]: E1203 13:36:24.651264 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="extract-content" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.651534 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="extract-content" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.651856 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930d8f5-e952-49d7-bcbe-02a39ee37267" containerName="registry-server" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.653725 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.673692 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k7rt9"] Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.853850 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-catalog-content\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.853909 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-utilities\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.854404 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzskg\" (UniqueName: \"kubernetes.io/projected/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-kube-api-access-jzskg\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.955938 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzskg\" (UniqueName: \"kubernetes.io/projected/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-kube-api-access-jzskg\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.956002 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-catalog-content\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.956019 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-utilities\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.956505 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-utilities\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.956707 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-catalog-content\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.977137 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzskg\" (UniqueName: \"kubernetes.io/projected/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-kube-api-access-jzskg\") pod \"certified-operators-k7rt9\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:24 crc kubenswrapper[4990]: I1203 13:36:24.978380 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:25 crc kubenswrapper[4990]: I1203 13:36:25.525957 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k7rt9"] Dec 03 13:36:25 crc kubenswrapper[4990]: I1203 13:36:25.532539 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7rt9" event={"ID":"eaafdfc1-8ca2-4387-902b-d4c63e6c3294","Type":"ContainerStarted","Data":"3fab179d034a1ce5d65bf01e02449f315de7530d834587465d1be9379234b4d5"} Dec 03 13:36:25 crc kubenswrapper[4990]: I1203 13:36:25.686060 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:26 crc kubenswrapper[4990]: I1203 13:36:26.544895 4990 generic.go:334] "Generic (PLEG): container finished" podID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerID="92ba5a3adc988b892d94cfb75c3583883b2d3fa62614f5582408b6e8da9482b3" exitCode=0 Dec 03 13:36:26 crc kubenswrapper[4990]: I1203 13:36:26.544951 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7rt9" event={"ID":"eaafdfc1-8ca2-4387-902b-d4c63e6c3294","Type":"ContainerDied","Data":"92ba5a3adc988b892d94cfb75c3583883b2d3fa62614f5582408b6e8da9482b3"} Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.009789 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f248l"] Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.010462 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f248l" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="registry-server" containerID="cri-o://69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9" gracePeriod=2 Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.496325 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.564252 4990 generic.go:334] "Generic (PLEG): container finished" podID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerID="69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9" exitCode=0 Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.564307 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f248l" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.564320 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f248l" event={"ID":"e81ddbb3-463e-4271-b0d6-baf34d74cb26","Type":"ContainerDied","Data":"69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9"} Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.564389 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f248l" event={"ID":"e81ddbb3-463e-4271-b0d6-baf34d74cb26","Type":"ContainerDied","Data":"715f51e04296bec766ea5c631494fc5d4fbe29ccccff0d5c77ae856305e5b4f7"} Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.564410 4990 scope.go:117] "RemoveContainer" containerID="69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.567282 4990 generic.go:334] "Generic (PLEG): container finished" podID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerID="7581ed7cba5a1fd7b5126646bd28506e6015d277751a2b08c7206dd7b06812be" exitCode=0 Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.567319 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7rt9" event={"ID":"eaafdfc1-8ca2-4387-902b-d4c63e6c3294","Type":"ContainerDied","Data":"7581ed7cba5a1fd7b5126646bd28506e6015d277751a2b08c7206dd7b06812be"} Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.568978 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.600540 4990 scope.go:117] "RemoveContainer" containerID="36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.622534 4990 scope.go:117] "RemoveContainer" containerID="1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.642913 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-utilities\") pod \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.643281 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcwsm\" (UniqueName: \"kubernetes.io/projected/e81ddbb3-463e-4271-b0d6-baf34d74cb26-kube-api-access-rcwsm\") pod \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.643352 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-utilities" (OuterVolumeSpecName: "utilities") pod "e81ddbb3-463e-4271-b0d6-baf34d74cb26" (UID: "e81ddbb3-463e-4271-b0d6-baf34d74cb26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.643379 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-catalog-content\") pod \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\" (UID: \"e81ddbb3-463e-4271-b0d6-baf34d74cb26\") " Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.647471 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.650397 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81ddbb3-463e-4271-b0d6-baf34d74cb26-kube-api-access-rcwsm" (OuterVolumeSpecName: "kube-api-access-rcwsm") pod "e81ddbb3-463e-4271-b0d6-baf34d74cb26" (UID: "e81ddbb3-463e-4271-b0d6-baf34d74cb26"). InnerVolumeSpecName "kube-api-access-rcwsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.670527 4990 scope.go:117] "RemoveContainer" containerID="69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9" Dec 03 13:36:28 crc kubenswrapper[4990]: E1203 13:36:28.673931 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9\": container with ID starting with 69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9 not found: ID does not exist" containerID="69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.673976 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9"} err="failed to get container status \"69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9\": rpc error: code = NotFound desc = could not find container \"69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9\": container with ID starting with 69af7cf8ee9f6a63d3eb55e400b28df58161ff8cf9e7fd1227b573012d6954f9 not found: ID does not exist" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.674000 4990 scope.go:117] "RemoveContainer" containerID="36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7" Dec 03 13:36:28 crc kubenswrapper[4990]: E1203 13:36:28.674369 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7\": container with ID starting with 36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7 not found: ID does not exist" containerID="36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.674391 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7"} err="failed to get container status \"36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7\": rpc error: code = NotFound desc = could not find container \"36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7\": container with ID starting with 36254e8da72427ac86ba952c6eaf1c6a659a19ddee7665b94a68036c0c067db7 not found: ID does not exist" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.674405 4990 scope.go:117] "RemoveContainer" containerID="1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0" Dec 03 13:36:28 crc kubenswrapper[4990]: E1203 13:36:28.674647 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0\": container with ID starting with 1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0 not found: ID does not exist" containerID="1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.674674 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0"} err="failed to get container status \"1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0\": rpc error: code = NotFound desc = could not find container \"1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0\": container with ID starting with 1f3de9a501e3cafb297a4c80ba2edbc503e06e5a187b4f6b20cfe822be1b77f0 not found: ID does not exist" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.687999 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e81ddbb3-463e-4271-b0d6-baf34d74cb26" (UID: "e81ddbb3-463e-4271-b0d6-baf34d74cb26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.749567 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81ddbb3-463e-4271-b0d6-baf34d74cb26-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.749619 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcwsm\" (UniqueName: \"kubernetes.io/projected/e81ddbb3-463e-4271-b0d6-baf34d74cb26-kube-api-access-rcwsm\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.918496 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f248l"] Dec 03 13:36:28 crc kubenswrapper[4990]: I1203 13:36:28.926961 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f248l"] Dec 03 13:36:29 crc kubenswrapper[4990]: I1203 13:36:29.578780 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7rt9" event={"ID":"eaafdfc1-8ca2-4387-902b-d4c63e6c3294","Type":"ContainerStarted","Data":"2104f1e5afb6c551616207a4e5eda725b57711ca8819311b3c98bce1d983f417"} Dec 03 13:36:29 crc kubenswrapper[4990]: I1203 13:36:29.597752 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k7rt9" podStartSLOduration=3.149823714 podStartE2EDuration="5.597736439s" podCreationTimestamp="2025-12-03 13:36:24 +0000 UTC" firstStartedPulling="2025-12-03 13:36:26.547222074 +0000 UTC m=+3534.689133293" lastFinishedPulling="2025-12-03 13:36:28.995134789 +0000 UTC m=+3537.137046018" observedRunningTime="2025-12-03 13:36:29.594045819 +0000 UTC m=+3537.735957048" watchObservedRunningTime="2025-12-03 13:36:29.597736439 +0000 UTC m=+3537.739647668" Dec 03 13:36:30 crc kubenswrapper[4990]: I1203 13:36:30.277094 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" path="/var/lib/kubelet/pods/e81ddbb3-463e-4271-b0d6-baf34d74cb26/volumes" Dec 03 13:36:34 crc kubenswrapper[4990]: I1203 13:36:34.979475 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:34 crc kubenswrapper[4990]: I1203 13:36:34.980010 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:35 crc kubenswrapper[4990]: I1203 13:36:35.031815 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:35 crc kubenswrapper[4990]: I1203 13:36:35.713115 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:37 crc kubenswrapper[4990]: I1203 13:36:37.012315 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k7rt9"] Dec 03 13:36:37 crc kubenswrapper[4990]: I1203 13:36:37.662970 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k7rt9" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="registry-server" containerID="cri-o://2104f1e5afb6c551616207a4e5eda725b57711ca8819311b3c98bce1d983f417" gracePeriod=2 Dec 03 13:36:38 crc kubenswrapper[4990]: I1203 13:36:38.677770 4990 generic.go:334] "Generic (PLEG): container finished" podID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerID="2104f1e5afb6c551616207a4e5eda725b57711ca8819311b3c98bce1d983f417" exitCode=0 Dec 03 13:36:38 crc kubenswrapper[4990]: I1203 13:36:38.677823 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7rt9" event={"ID":"eaafdfc1-8ca2-4387-902b-d4c63e6c3294","Type":"ContainerDied","Data":"2104f1e5afb6c551616207a4e5eda725b57711ca8819311b3c98bce1d983f417"} Dec 03 13:36:38 crc kubenswrapper[4990]: I1203 13:36:38.970912 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.089271 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-catalog-content\") pod \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.089663 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-utilities\") pod \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.089700 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzskg\" (UniqueName: \"kubernetes.io/projected/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-kube-api-access-jzskg\") pod \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\" (UID: \"eaafdfc1-8ca2-4387-902b-d4c63e6c3294\") " Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.090397 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-utilities" (OuterVolumeSpecName: "utilities") pod "eaafdfc1-8ca2-4387-902b-d4c63e6c3294" (UID: "eaafdfc1-8ca2-4387-902b-d4c63e6c3294"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.105729 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-kube-api-access-jzskg" (OuterVolumeSpecName: "kube-api-access-jzskg") pod "eaafdfc1-8ca2-4387-902b-d4c63e6c3294" (UID: "eaafdfc1-8ca2-4387-902b-d4c63e6c3294"). InnerVolumeSpecName "kube-api-access-jzskg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.155111 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eaafdfc1-8ca2-4387-902b-d4c63e6c3294" (UID: "eaafdfc1-8ca2-4387-902b-d4c63e6c3294"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.192036 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.192073 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzskg\" (UniqueName: \"kubernetes.io/projected/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-kube-api-access-jzskg\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.192089 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaafdfc1-8ca2-4387-902b-d4c63e6c3294-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.691785 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7rt9" event={"ID":"eaafdfc1-8ca2-4387-902b-d4c63e6c3294","Type":"ContainerDied","Data":"3fab179d034a1ce5d65bf01e02449f315de7530d834587465d1be9379234b4d5"} Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.692107 4990 scope.go:117] "RemoveContainer" containerID="2104f1e5afb6c551616207a4e5eda725b57711ca8819311b3c98bce1d983f417" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.691876 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7rt9" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.725603 4990 scope.go:117] "RemoveContainer" containerID="7581ed7cba5a1fd7b5126646bd28506e6015d277751a2b08c7206dd7b06812be" Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.740625 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k7rt9"] Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.750964 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k7rt9"] Dec 03 13:36:39 crc kubenswrapper[4990]: I1203 13:36:39.768477 4990 scope.go:117] "RemoveContainer" containerID="92ba5a3adc988b892d94cfb75c3583883b2d3fa62614f5582408b6e8da9482b3" Dec 03 13:36:40 crc kubenswrapper[4990]: I1203 13:36:40.279703 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" path="/var/lib/kubelet/pods/eaafdfc1-8ca2-4387-902b-d4c63e6c3294/volumes" Dec 03 13:38:03 crc kubenswrapper[4990]: I1203 13:38:03.288788 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:38:03 crc kubenswrapper[4990]: I1203 13:38:03.289802 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:38:33 crc kubenswrapper[4990]: I1203 13:38:33.286251 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:38:33 crc kubenswrapper[4990]: I1203 13:38:33.286776 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:39:03 crc kubenswrapper[4990]: I1203 13:39:03.287054 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:39:03 crc kubenswrapper[4990]: I1203 13:39:03.287808 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:39:03 crc kubenswrapper[4990]: I1203 13:39:03.287889 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:39:03 crc kubenswrapper[4990]: I1203 13:39:03.289029 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:39:03 crc kubenswrapper[4990]: I1203 13:39:03.289136 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" gracePeriod=600 Dec 03 13:39:03 crc kubenswrapper[4990]: E1203 13:39:03.407568 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:39:04 crc kubenswrapper[4990]: I1203 13:39:04.073505 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" exitCode=0 Dec 03 13:39:04 crc kubenswrapper[4990]: I1203 13:39:04.073583 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173"} Dec 03 13:39:04 crc kubenswrapper[4990]: I1203 13:39:04.073884 4990 scope.go:117] "RemoveContainer" containerID="759279b6d38b81a7f11d8d2f79cf938abb09838e4952a8a80b89c5ee5b26fce1" Dec 03 13:39:04 crc kubenswrapper[4990]: I1203 13:39:04.074765 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:39:04 crc kubenswrapper[4990]: E1203 13:39:04.075139 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:39:18 crc kubenswrapper[4990]: I1203 13:39:18.265252 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:39:18 crc kubenswrapper[4990]: E1203 13:39:18.266042 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:39:30 crc kubenswrapper[4990]: I1203 13:39:30.264346 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:39:30 crc kubenswrapper[4990]: E1203 13:39:30.265149 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:39:41 crc kubenswrapper[4990]: I1203 13:39:41.265347 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:39:41 crc kubenswrapper[4990]: E1203 13:39:41.266228 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:39:54 crc kubenswrapper[4990]: I1203 13:39:54.265578 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:39:54 crc kubenswrapper[4990]: E1203 13:39:54.266386 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:40:07 crc kubenswrapper[4990]: I1203 13:40:07.264871 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:40:07 crc kubenswrapper[4990]: E1203 13:40:07.265873 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:40:21 crc kubenswrapper[4990]: I1203 13:40:21.265320 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:40:21 crc kubenswrapper[4990]: E1203 13:40:21.266370 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:40:33 crc kubenswrapper[4990]: I1203 13:40:33.265428 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:40:33 crc kubenswrapper[4990]: E1203 13:40:33.266618 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:40:48 crc kubenswrapper[4990]: I1203 13:40:48.268091 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:40:48 crc kubenswrapper[4990]: E1203 13:40:48.268900 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:41:01 crc kubenswrapper[4990]: I1203 13:41:01.264565 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:41:01 crc kubenswrapper[4990]: E1203 13:41:01.265382 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:41:15 crc kubenswrapper[4990]: I1203 13:41:15.265228 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:41:15 crc kubenswrapper[4990]: E1203 13:41:15.266100 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:41:27 crc kubenswrapper[4990]: I1203 13:41:27.264716 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:41:27 crc kubenswrapper[4990]: E1203 13:41:27.265491 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:41:39 crc kubenswrapper[4990]: I1203 13:41:39.265065 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:41:39 crc kubenswrapper[4990]: E1203 13:41:39.265903 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:41:54 crc kubenswrapper[4990]: I1203 13:41:54.265145 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:41:54 crc kubenswrapper[4990]: E1203 13:41:54.266025 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:42:08 crc kubenswrapper[4990]: I1203 13:42:08.264247 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:42:08 crc kubenswrapper[4990]: E1203 13:42:08.265135 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:42:21 crc kubenswrapper[4990]: I1203 13:42:21.264879 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:42:21 crc kubenswrapper[4990]: E1203 13:42:21.265740 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:42:34 crc kubenswrapper[4990]: I1203 13:42:34.264643 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:42:34 crc kubenswrapper[4990]: E1203 13:42:34.265369 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.032874 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tnklt"] Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.033997 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="registry-server" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034013 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="registry-server" Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.034037 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="extract-content" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034044 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="extract-content" Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.034060 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="registry-server" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034067 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="registry-server" Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.034097 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="extract-content" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034104 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="extract-content" Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.034120 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="extract-utilities" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034127 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="extract-utilities" Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.034137 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="extract-utilities" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034144 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="extract-utilities" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034369 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaafdfc1-8ca2-4387-902b-d4c63e6c3294" containerName="registry-server" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.034392 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81ddbb3-463e-4271-b0d6-baf34d74cb26" containerName="registry-server" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.036007 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.051985 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tnklt"] Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.142119 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-utilities\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.142195 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zh2z\" (UniqueName: \"kubernetes.io/projected/04d98c58-c05e-4165-8665-0757ef383fd2-kube-api-access-2zh2z\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.142335 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-catalog-content\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.244139 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-catalog-content\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.244305 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-utilities\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.244355 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zh2z\" (UniqueName: \"kubernetes.io/projected/04d98c58-c05e-4165-8665-0757ef383fd2-kube-api-access-2zh2z\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.244842 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-catalog-content\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.244885 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-utilities\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.264937 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:42:48 crc kubenswrapper[4990]: E1203 13:42:48.265496 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.270582 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zh2z\" (UniqueName: \"kubernetes.io/projected/04d98c58-c05e-4165-8665-0757ef383fd2-kube-api-access-2zh2z\") pod \"redhat-marketplace-tnklt\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.366243 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:48 crc kubenswrapper[4990]: I1203 13:42:48.898201 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tnklt"] Dec 03 13:42:49 crc kubenswrapper[4990]: W1203 13:42:49.192288 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04d98c58_c05e_4165_8665_0757ef383fd2.slice/crio-f48e08c809e8fa0b9ea6f1f078f5442c1a1fdd43c0246a763da1b3187987717c WatchSource:0}: Error finding container f48e08c809e8fa0b9ea6f1f078f5442c1a1fdd43c0246a763da1b3187987717c: Status 404 returned error can't find the container with id f48e08c809e8fa0b9ea6f1f078f5442c1a1fdd43c0246a763da1b3187987717c Dec 03 13:42:49 crc kubenswrapper[4990]: I1203 13:42:49.658541 4990 generic.go:334] "Generic (PLEG): container finished" podID="04d98c58-c05e-4165-8665-0757ef383fd2" containerID="ee5bd0dd67b87caa22aedbf58cfc5256e843e3f210c03bfb08260b720f532c5f" exitCode=0 Dec 03 13:42:49 crc kubenswrapper[4990]: I1203 13:42:49.658598 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerDied","Data":"ee5bd0dd67b87caa22aedbf58cfc5256e843e3f210c03bfb08260b720f532c5f"} Dec 03 13:42:49 crc kubenswrapper[4990]: I1203 13:42:49.658903 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerStarted","Data":"f48e08c809e8fa0b9ea6f1f078f5442c1a1fdd43c0246a763da1b3187987717c"} Dec 03 13:42:49 crc kubenswrapper[4990]: I1203 13:42:49.661229 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:42:50 crc kubenswrapper[4990]: I1203 13:42:50.668971 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerStarted","Data":"cfed5ebbd47cc23cb6e64698a533adaba50880d9ca8b60efe9770541dc4e3e0d"} Dec 03 13:42:51 crc kubenswrapper[4990]: I1203 13:42:51.682265 4990 generic.go:334] "Generic (PLEG): container finished" podID="04d98c58-c05e-4165-8665-0757ef383fd2" containerID="cfed5ebbd47cc23cb6e64698a533adaba50880d9ca8b60efe9770541dc4e3e0d" exitCode=0 Dec 03 13:42:51 crc kubenswrapper[4990]: I1203 13:42:51.682336 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerDied","Data":"cfed5ebbd47cc23cb6e64698a533adaba50880d9ca8b60efe9770541dc4e3e0d"} Dec 03 13:42:52 crc kubenswrapper[4990]: I1203 13:42:52.709963 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerStarted","Data":"04badae7c8d392d6974b50394c9b509aef3eb6a9aa668f3b5baaa241cd872500"} Dec 03 13:42:58 crc kubenswrapper[4990]: I1203 13:42:58.366815 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:58 crc kubenswrapper[4990]: I1203 13:42:58.367314 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:58 crc kubenswrapper[4990]: I1203 13:42:58.422678 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:58 crc kubenswrapper[4990]: I1203 13:42:58.446737 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tnklt" podStartSLOduration=8.023112144 podStartE2EDuration="10.446719451s" podCreationTimestamp="2025-12-03 13:42:48 +0000 UTC" firstStartedPulling="2025-12-03 13:42:49.660836748 +0000 UTC m=+3917.802747977" lastFinishedPulling="2025-12-03 13:42:52.084444055 +0000 UTC m=+3920.226355284" observedRunningTime="2025-12-03 13:42:52.73230495 +0000 UTC m=+3920.874216179" watchObservedRunningTime="2025-12-03 13:42:58.446719451 +0000 UTC m=+3926.588630680" Dec 03 13:42:59 crc kubenswrapper[4990]: I1203 13:42:59.442851 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:42:59 crc kubenswrapper[4990]: I1203 13:42:59.509190 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tnklt"] Dec 03 13:43:00 crc kubenswrapper[4990]: I1203 13:43:00.774572 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tnklt" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="registry-server" containerID="cri-o://04badae7c8d392d6974b50394c9b509aef3eb6a9aa668f3b5baaa241cd872500" gracePeriod=2 Dec 03 13:43:01 crc kubenswrapper[4990]: I1203 13:43:01.264279 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:43:01 crc kubenswrapper[4990]: E1203 13:43:01.264667 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:43:01 crc kubenswrapper[4990]: I1203 13:43:01.788666 4990 generic.go:334] "Generic (PLEG): container finished" podID="04d98c58-c05e-4165-8665-0757ef383fd2" containerID="04badae7c8d392d6974b50394c9b509aef3eb6a9aa668f3b5baaa241cd872500" exitCode=0 Dec 03 13:43:01 crc kubenswrapper[4990]: I1203 13:43:01.788739 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerDied","Data":"04badae7c8d392d6974b50394c9b509aef3eb6a9aa668f3b5baaa241cd872500"} Dec 03 13:43:01 crc kubenswrapper[4990]: I1203 13:43:01.789015 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tnklt" event={"ID":"04d98c58-c05e-4165-8665-0757ef383fd2","Type":"ContainerDied","Data":"f48e08c809e8fa0b9ea6f1f078f5442c1a1fdd43c0246a763da1b3187987717c"} Dec 03 13:43:01 crc kubenswrapper[4990]: I1203 13:43:01.789034 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f48e08c809e8fa0b9ea6f1f078f5442c1a1fdd43c0246a763da1b3187987717c" Dec 03 13:43:01 crc kubenswrapper[4990]: I1203 13:43:01.876216 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.012146 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-catalog-content\") pod \"04d98c58-c05e-4165-8665-0757ef383fd2\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.012229 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-utilities\") pod \"04d98c58-c05e-4165-8665-0757ef383fd2\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.012483 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zh2z\" (UniqueName: \"kubernetes.io/projected/04d98c58-c05e-4165-8665-0757ef383fd2-kube-api-access-2zh2z\") pod \"04d98c58-c05e-4165-8665-0757ef383fd2\" (UID: \"04d98c58-c05e-4165-8665-0757ef383fd2\") " Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.013253 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-utilities" (OuterVolumeSpecName: "utilities") pod "04d98c58-c05e-4165-8665-0757ef383fd2" (UID: "04d98c58-c05e-4165-8665-0757ef383fd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.023847 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d98c58-c05e-4165-8665-0757ef383fd2-kube-api-access-2zh2z" (OuterVolumeSpecName: "kube-api-access-2zh2z") pod "04d98c58-c05e-4165-8665-0757ef383fd2" (UID: "04d98c58-c05e-4165-8665-0757ef383fd2"). InnerVolumeSpecName "kube-api-access-2zh2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.033877 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04d98c58-c05e-4165-8665-0757ef383fd2" (UID: "04d98c58-c05e-4165-8665-0757ef383fd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.114311 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.114664 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zh2z\" (UniqueName: \"kubernetes.io/projected/04d98c58-c05e-4165-8665-0757ef383fd2-kube-api-access-2zh2z\") on node \"crc\" DevicePath \"\"" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.114674 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d98c58-c05e-4165-8665-0757ef383fd2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.798283 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tnklt" Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.827375 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tnklt"] Dec 03 13:43:02 crc kubenswrapper[4990]: I1203 13:43:02.837998 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tnklt"] Dec 03 13:43:04 crc kubenswrapper[4990]: I1203 13:43:04.276758 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" path="/var/lib/kubelet/pods/04d98c58-c05e-4165-8665-0757ef383fd2/volumes" Dec 03 13:43:16 crc kubenswrapper[4990]: I1203 13:43:16.267798 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:43:16 crc kubenswrapper[4990]: E1203 13:43:16.269822 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:43:27 crc kubenswrapper[4990]: I1203 13:43:27.264967 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:43:27 crc kubenswrapper[4990]: E1203 13:43:27.265694 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:43:39 crc kubenswrapper[4990]: I1203 13:43:39.264865 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:43:39 crc kubenswrapper[4990]: E1203 13:43:39.265795 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:43:50 crc kubenswrapper[4990]: I1203 13:43:50.264777 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:43:50 crc kubenswrapper[4990]: E1203 13:43:50.265557 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:44:04 crc kubenswrapper[4990]: I1203 13:44:04.264276 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:44:05 crc kubenswrapper[4990]: I1203 13:44:05.411440 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"33931fd5dcd79ed868236bde132bdc93fe9fd197d027f2241b26f817e20a5e50"} Dec 03 13:44:48 crc kubenswrapper[4990]: I1203 13:44:48.790880 4990 generic.go:334] "Generic (PLEG): container finished" podID="f375426e-6095-4621-ab29-6c7324da86e7" containerID="630e02c903627ad97e6cc587cad2680a000e819c3908ebda0e544b5ef2fcf983" exitCode=0 Dec 03 13:44:48 crc kubenswrapper[4990]: I1203 13:44:48.791129 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f375426e-6095-4621-ab29-6c7324da86e7","Type":"ContainerDied","Data":"630e02c903627ad97e6cc587cad2680a000e819c3908ebda0e544b5ef2fcf983"} Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.297981 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471470 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-workdir\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471610 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471629 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ssh-key\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471646 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-config-data\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471686 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ca-certs\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471708 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config-secret\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471743 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471825 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-temporary\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.471854 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c68bt\" (UniqueName: \"kubernetes.io/projected/f375426e-6095-4621-ab29-6c7324da86e7-kube-api-access-c68bt\") pod \"f375426e-6095-4621-ab29-6c7324da86e7\" (UID: \"f375426e-6095-4621-ab29-6c7324da86e7\") " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.472861 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-config-data" (OuterVolumeSpecName: "config-data") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.473421 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.476075 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.478199 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f375426e-6095-4621-ab29-6c7324da86e7-kube-api-access-c68bt" (OuterVolumeSpecName: "kube-api-access-c68bt") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "kube-api-access-c68bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.480152 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.503778 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.506216 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.509635 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.525234 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f375426e-6095-4621-ab29-6c7324da86e7" (UID: "f375426e-6095-4621-ab29-6c7324da86e7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574109 4990 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574148 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c68bt\" (UniqueName: \"kubernetes.io/projected/f375426e-6095-4621-ab29-6c7324da86e7-kube-api-access-c68bt\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574164 4990 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f375426e-6095-4621-ab29-6c7324da86e7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574177 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574190 4990 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574203 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f375426e-6095-4621-ab29-6c7324da86e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574213 4990 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574223 4990 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f375426e-6095-4621-ab29-6c7324da86e7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.574259 4990 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.605219 4990 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.676119 4990 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.812283 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"f375426e-6095-4621-ab29-6c7324da86e7","Type":"ContainerDied","Data":"ce20b67ff8e6e9c2a5f66d2043554b636d5074adafd43a65ce7891b54fdc56c9"} Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.812327 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce20b67ff8e6e9c2a5f66d2043554b636d5074adafd43a65ce7891b54fdc56c9" Dec 03 13:44:50 crc kubenswrapper[4990]: I1203 13:44:50.812356 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.963486 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 13:44:56 crc kubenswrapper[4990]: E1203 13:44:56.965725 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="extract-utilities" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.965763 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="extract-utilities" Dec 03 13:44:56 crc kubenswrapper[4990]: E1203 13:44:56.965780 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="registry-server" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.965788 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="registry-server" Dec 03 13:44:56 crc kubenswrapper[4990]: E1203 13:44:56.965802 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="extract-content" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.965808 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="extract-content" Dec 03 13:44:56 crc kubenswrapper[4990]: E1203 13:44:56.965822 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f375426e-6095-4621-ab29-6c7324da86e7" containerName="tempest-tests-tempest-tests-runner" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.965827 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f375426e-6095-4621-ab29-6c7324da86e7" containerName="tempest-tests-tempest-tests-runner" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.966013 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d98c58-c05e-4165-8665-0757ef383fd2" containerName="registry-server" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.966035 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f375426e-6095-4621-ab29-6c7324da86e7" containerName="tempest-tests-tempest-tests-runner" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.966683 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.968940 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-c2cfh" Dec 03 13:44:56 crc kubenswrapper[4990]: I1203 13:44:56.973947 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.093802 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.094030 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xknj2\" (UniqueName: \"kubernetes.io/projected/f0fab240-b45f-44e1-8fb7-446c4dc6d64e-kube-api-access-xknj2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.196010 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xknj2\" (UniqueName: \"kubernetes.io/projected/f0fab240-b45f-44e1-8fb7-446c4dc6d64e-kube-api-access-xknj2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.196474 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.197233 4990 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.220388 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xknj2\" (UniqueName: \"kubernetes.io/projected/f0fab240-b45f-44e1-8fb7-446c4dc6d64e-kube-api-access-xknj2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.226196 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f0fab240-b45f-44e1-8fb7-446c4dc6d64e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.286387 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.733006 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 13:44:57 crc kubenswrapper[4990]: I1203 13:44:57.874030 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f0fab240-b45f-44e1-8fb7-446c4dc6d64e","Type":"ContainerStarted","Data":"c17d64e157aded54915d2398de62302318bdb489a28cacb14d66a3d83e3bf141"} Dec 03 13:44:59 crc kubenswrapper[4990]: I1203 13:44:59.892315 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f0fab240-b45f-44e1-8fb7-446c4dc6d64e","Type":"ContainerStarted","Data":"78f115e488c56fecde10332649bcec9728aaaddcac415a571910107c8eb2dcb0"} Dec 03 13:44:59 crc kubenswrapper[4990]: I1203 13:44:59.920033 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.480167421 podStartE2EDuration="3.920010596s" podCreationTimestamp="2025-12-03 13:44:56 +0000 UTC" firstStartedPulling="2025-12-03 13:44:57.740395672 +0000 UTC m=+4045.882306891" lastFinishedPulling="2025-12-03 13:44:59.180238837 +0000 UTC m=+4047.322150066" observedRunningTime="2025-12-03 13:44:59.906118057 +0000 UTC m=+4048.048029316" watchObservedRunningTime="2025-12-03 13:44:59.920010596 +0000 UTC m=+4048.061921835" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.172316 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf"] Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.173705 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.175932 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.176061 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.183782 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf"] Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.257787 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgg8t\" (UniqueName: \"kubernetes.io/projected/a532916c-5b07-4d65-841f-bc0767b0931e-kube-api-access-lgg8t\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.258215 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a532916c-5b07-4d65-841f-bc0767b0931e-secret-volume\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.258360 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a532916c-5b07-4d65-841f-bc0767b0931e-config-volume\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.360201 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a532916c-5b07-4d65-841f-bc0767b0931e-secret-volume\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.360297 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a532916c-5b07-4d65-841f-bc0767b0931e-config-volume\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.360394 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgg8t\" (UniqueName: \"kubernetes.io/projected/a532916c-5b07-4d65-841f-bc0767b0931e-kube-api-access-lgg8t\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.361314 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a532916c-5b07-4d65-841f-bc0767b0931e-config-volume\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.366800 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a532916c-5b07-4d65-841f-bc0767b0931e-secret-volume\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.379402 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgg8t\" (UniqueName: \"kubernetes.io/projected/a532916c-5b07-4d65-841f-bc0767b0931e-kube-api-access-lgg8t\") pod \"collect-profiles-29412825-5b2tf\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.491538 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:00 crc kubenswrapper[4990]: I1203 13:45:00.944421 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf"] Dec 03 13:45:00 crc kubenswrapper[4990]: W1203 13:45:00.945161 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda532916c_5b07_4d65_841f_bc0767b0931e.slice/crio-36c4eecfb8bf6ec2fd7ee3b70bdd105160f1cc8022ba37259fbd1f6dbed56b68 WatchSource:0}: Error finding container 36c4eecfb8bf6ec2fd7ee3b70bdd105160f1cc8022ba37259fbd1f6dbed56b68: Status 404 returned error can't find the container with id 36c4eecfb8bf6ec2fd7ee3b70bdd105160f1cc8022ba37259fbd1f6dbed56b68 Dec 03 13:45:01 crc kubenswrapper[4990]: I1203 13:45:01.914780 4990 generic.go:334] "Generic (PLEG): container finished" podID="a532916c-5b07-4d65-841f-bc0767b0931e" containerID="ceb3e62e2bbfd7241730111e06b7945590991afd0bd2b32320bbb3e7b317e4e3" exitCode=0 Dec 03 13:45:01 crc kubenswrapper[4990]: I1203 13:45:01.914937 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" event={"ID":"a532916c-5b07-4d65-841f-bc0767b0931e","Type":"ContainerDied","Data":"ceb3e62e2bbfd7241730111e06b7945590991afd0bd2b32320bbb3e7b317e4e3"} Dec 03 13:45:01 crc kubenswrapper[4990]: I1203 13:45:01.915536 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" event={"ID":"a532916c-5b07-4d65-841f-bc0767b0931e","Type":"ContainerStarted","Data":"36c4eecfb8bf6ec2fd7ee3b70bdd105160f1cc8022ba37259fbd1f6dbed56b68"} Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.349292 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.523159 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a532916c-5b07-4d65-841f-bc0767b0931e-config-volume\") pod \"a532916c-5b07-4d65-841f-bc0767b0931e\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.523269 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a532916c-5b07-4d65-841f-bc0767b0931e-secret-volume\") pod \"a532916c-5b07-4d65-841f-bc0767b0931e\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.523440 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgg8t\" (UniqueName: \"kubernetes.io/projected/a532916c-5b07-4d65-841f-bc0767b0931e-kube-api-access-lgg8t\") pod \"a532916c-5b07-4d65-841f-bc0767b0931e\" (UID: \"a532916c-5b07-4d65-841f-bc0767b0931e\") " Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.524123 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a532916c-5b07-4d65-841f-bc0767b0931e-config-volume" (OuterVolumeSpecName: "config-volume") pod "a532916c-5b07-4d65-841f-bc0767b0931e" (UID: "a532916c-5b07-4d65-841f-bc0767b0931e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.524230 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a532916c-5b07-4d65-841f-bc0767b0931e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.529434 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a532916c-5b07-4d65-841f-bc0767b0931e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a532916c-5b07-4d65-841f-bc0767b0931e" (UID: "a532916c-5b07-4d65-841f-bc0767b0931e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.530052 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a532916c-5b07-4d65-841f-bc0767b0931e-kube-api-access-lgg8t" (OuterVolumeSpecName: "kube-api-access-lgg8t") pod "a532916c-5b07-4d65-841f-bc0767b0931e" (UID: "a532916c-5b07-4d65-841f-bc0767b0931e"). InnerVolumeSpecName "kube-api-access-lgg8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.626053 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a532916c-5b07-4d65-841f-bc0767b0931e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.626109 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgg8t\" (UniqueName: \"kubernetes.io/projected/a532916c-5b07-4d65-841f-bc0767b0931e-kube-api-access-lgg8t\") on node \"crc\" DevicePath \"\"" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.935001 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" event={"ID":"a532916c-5b07-4d65-841f-bc0767b0931e","Type":"ContainerDied","Data":"36c4eecfb8bf6ec2fd7ee3b70bdd105160f1cc8022ba37259fbd1f6dbed56b68"} Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.935420 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36c4eecfb8bf6ec2fd7ee3b70bdd105160f1cc8022ba37259fbd1f6dbed56b68" Dec 03 13:45:03 crc kubenswrapper[4990]: I1203 13:45:03.935063 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412825-5b2tf" Dec 03 13:45:04 crc kubenswrapper[4990]: I1203 13:45:04.424214 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc"] Dec 03 13:45:04 crc kubenswrapper[4990]: I1203 13:45:04.431254 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412780-vrpqc"] Dec 03 13:45:06 crc kubenswrapper[4990]: I1203 13:45:06.278470 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f73e432-c318-45bf-a8ec-93b26f1f823d" path="/var/lib/kubelet/pods/7f73e432-c318-45bf-a8ec-93b26f1f823d/volumes" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.188185 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q8ss7/must-gather-fd98h"] Dec 03 13:45:24 crc kubenswrapper[4990]: E1203 13:45:24.189364 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a532916c-5b07-4d65-841f-bc0767b0931e" containerName="collect-profiles" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.189384 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="a532916c-5b07-4d65-841f-bc0767b0931e" containerName="collect-profiles" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.189746 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="a532916c-5b07-4d65-841f-bc0767b0931e" containerName="collect-profiles" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.190972 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.194477 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-q8ss7"/"default-dockercfg-qd5rs" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.194693 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-q8ss7"/"openshift-service-ca.crt" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.194901 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-q8ss7"/"kube-root-ca.crt" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.222026 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-q8ss7/must-gather-fd98h"] Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.362189 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mj8t\" (UniqueName: \"kubernetes.io/projected/741c2504-9580-4514-bf3a-0de631f99c03-kube-api-access-4mj8t\") pod \"must-gather-fd98h\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.362331 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/741c2504-9580-4514-bf3a-0de631f99c03-must-gather-output\") pod \"must-gather-fd98h\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.463854 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mj8t\" (UniqueName: \"kubernetes.io/projected/741c2504-9580-4514-bf3a-0de631f99c03-kube-api-access-4mj8t\") pod \"must-gather-fd98h\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.463985 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/741c2504-9580-4514-bf3a-0de631f99c03-must-gather-output\") pod \"must-gather-fd98h\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.464473 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/741c2504-9580-4514-bf3a-0de631f99c03-must-gather-output\") pod \"must-gather-fd98h\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.486201 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mj8t\" (UniqueName: \"kubernetes.io/projected/741c2504-9580-4514-bf3a-0de631f99c03-kube-api-access-4mj8t\") pod \"must-gather-fd98h\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.531148 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:45:24 crc kubenswrapper[4990]: I1203 13:45:24.980250 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-q8ss7/must-gather-fd98h"] Dec 03 13:45:25 crc kubenswrapper[4990]: I1203 13:45:25.146134 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/must-gather-fd98h" event={"ID":"741c2504-9580-4514-bf3a-0de631f99c03","Type":"ContainerStarted","Data":"b1979103c96ec3dee4128be8f9b2c8ff6ed93184ea37370f909121d7165249d8"} Dec 03 13:45:26 crc kubenswrapper[4990]: I1203 13:45:26.942630 4990 scope.go:117] "RemoveContainer" containerID="5f8ce297f026d7e58d417b5143bfebe67c674959542e044a98564069411d56f0" Dec 03 13:45:34 crc kubenswrapper[4990]: I1203 13:45:34.244704 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/must-gather-fd98h" event={"ID":"741c2504-9580-4514-bf3a-0de631f99c03","Type":"ContainerStarted","Data":"ab6b98faaa61f3d31fd4cd7cc909fbeb310c5a75efdeb6291f6a98ee258851ba"} Dec 03 13:45:35 crc kubenswrapper[4990]: I1203 13:45:35.255829 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/must-gather-fd98h" event={"ID":"741c2504-9580-4514-bf3a-0de631f99c03","Type":"ContainerStarted","Data":"51ca4fa7d11dcbc0f1159614a31c6413d6a421f99631397e2dcc509b856fa33e"} Dec 03 13:45:35 crc kubenswrapper[4990]: I1203 13:45:35.275504 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q8ss7/must-gather-fd98h" podStartSLOduration=2.952652116 podStartE2EDuration="11.275486035s" podCreationTimestamp="2025-12-03 13:45:24 +0000 UTC" firstStartedPulling="2025-12-03 13:45:24.987039597 +0000 UTC m=+4073.128950826" lastFinishedPulling="2025-12-03 13:45:33.309873496 +0000 UTC m=+4081.451784745" observedRunningTime="2025-12-03 13:45:35.271689051 +0000 UTC m=+4083.413600300" watchObservedRunningTime="2025-12-03 13:45:35.275486035 +0000 UTC m=+4083.417397264" Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.640895 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-gjpjf"] Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.642218 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.788563 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa866d7b-f153-4962-a751-c2744d06b165-host\") pod \"crc-debug-gjpjf\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.789035 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twwdv\" (UniqueName: \"kubernetes.io/projected/aa866d7b-f153-4962-a751-c2744d06b165-kube-api-access-twwdv\") pod \"crc-debug-gjpjf\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.891222 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twwdv\" (UniqueName: \"kubernetes.io/projected/aa866d7b-f153-4962-a751-c2744d06b165-kube-api-access-twwdv\") pod \"crc-debug-gjpjf\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.891386 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa866d7b-f153-4962-a751-c2744d06b165-host\") pod \"crc-debug-gjpjf\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:37 crc kubenswrapper[4990]: I1203 13:45:37.891561 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa866d7b-f153-4962-a751-c2744d06b165-host\") pod \"crc-debug-gjpjf\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:38 crc kubenswrapper[4990]: I1203 13:45:38.381873 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twwdv\" (UniqueName: \"kubernetes.io/projected/aa866d7b-f153-4962-a751-c2744d06b165-kube-api-access-twwdv\") pod \"crc-debug-gjpjf\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:38 crc kubenswrapper[4990]: I1203 13:45:38.558670 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:45:39 crc kubenswrapper[4990]: I1203 13:45:39.298422 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" event={"ID":"aa866d7b-f153-4962-a751-c2744d06b165","Type":"ContainerStarted","Data":"ffc7a8ea362298dfcb481cbb564bc669a423026a454e2ebb8932cb304e279e39"} Dec 03 13:45:51 crc kubenswrapper[4990]: I1203 13:45:51.439119 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" event={"ID":"aa866d7b-f153-4962-a751-c2744d06b165","Type":"ContainerStarted","Data":"904e528b2f302ad44050994e6c6aac295215a3d24aae9530ab47dee372a02927"} Dec 03 13:45:51 crc kubenswrapper[4990]: I1203 13:45:51.459014 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" podStartSLOduration=2.519408635 podStartE2EDuration="14.45899422s" podCreationTimestamp="2025-12-03 13:45:37 +0000 UTC" firstStartedPulling="2025-12-03 13:45:38.592489912 +0000 UTC m=+4086.734401141" lastFinishedPulling="2025-12-03 13:45:50.532075487 +0000 UTC m=+4098.673986726" observedRunningTime="2025-12-03 13:45:51.457598602 +0000 UTC m=+4099.599509831" watchObservedRunningTime="2025-12-03 13:45:51.45899422 +0000 UTC m=+4099.600905449" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.365534 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fhv9c"] Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.370962 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.418786 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fhv9c"] Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.476585 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-catalog-content\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.476683 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kflhz\" (UniqueName: \"kubernetes.io/projected/d31b4fa7-9f88-490d-9b56-285739156f99-kube-api-access-kflhz\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.476751 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-utilities\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.578997 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kflhz\" (UniqueName: \"kubernetes.io/projected/d31b4fa7-9f88-490d-9b56-285739156f99-kube-api-access-kflhz\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.579096 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-utilities\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.579198 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-catalog-content\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.579675 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-utilities\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.579800 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-catalog-content\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:20 crc kubenswrapper[4990]: I1203 13:46:20.980902 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kflhz\" (UniqueName: \"kubernetes.io/projected/d31b4fa7-9f88-490d-9b56-285739156f99-kube-api-access-kflhz\") pod \"community-operators-fhv9c\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:21 crc kubenswrapper[4990]: I1203 13:46:21.001629 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:21 crc kubenswrapper[4990]: I1203 13:46:21.535486 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fhv9c"] Dec 03 13:46:21 crc kubenswrapper[4990]: I1203 13:46:21.738628 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhv9c" event={"ID":"d31b4fa7-9f88-490d-9b56-285739156f99","Type":"ContainerStarted","Data":"eb996dfa971c20a022f6cde749035605747bad0d865d4ee7c483ad971d974688"} Dec 03 13:46:22 crc kubenswrapper[4990]: I1203 13:46:22.747905 4990 generic.go:334] "Generic (PLEG): container finished" podID="d31b4fa7-9f88-490d-9b56-285739156f99" containerID="e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7" exitCode=0 Dec 03 13:46:22 crc kubenswrapper[4990]: I1203 13:46:22.748150 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhv9c" event={"ID":"d31b4fa7-9f88-490d-9b56-285739156f99","Type":"ContainerDied","Data":"e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7"} Dec 03 13:46:24 crc kubenswrapper[4990]: I1203 13:46:24.766386 4990 generic.go:334] "Generic (PLEG): container finished" podID="d31b4fa7-9f88-490d-9b56-285739156f99" containerID="8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35" exitCode=0 Dec 03 13:46:24 crc kubenswrapper[4990]: I1203 13:46:24.766525 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhv9c" event={"ID":"d31b4fa7-9f88-490d-9b56-285739156f99","Type":"ContainerDied","Data":"8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35"} Dec 03 13:46:26 crc kubenswrapper[4990]: I1203 13:46:26.788121 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhv9c" event={"ID":"d31b4fa7-9f88-490d-9b56-285739156f99","Type":"ContainerStarted","Data":"9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82"} Dec 03 13:46:26 crc kubenswrapper[4990]: I1203 13:46:26.810318 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fhv9c" podStartSLOduration=3.7431495630000002 podStartE2EDuration="6.810300219s" podCreationTimestamp="2025-12-03 13:46:20 +0000 UTC" firstStartedPulling="2025-12-03 13:46:22.750059979 +0000 UTC m=+4130.891971208" lastFinishedPulling="2025-12-03 13:46:25.817210615 +0000 UTC m=+4133.959121864" observedRunningTime="2025-12-03 13:46:26.80663266 +0000 UTC m=+4134.948543889" watchObservedRunningTime="2025-12-03 13:46:26.810300219 +0000 UTC m=+4134.952211478" Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.044410 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.045055 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.114109 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.830344 4990 generic.go:334] "Generic (PLEG): container finished" podID="aa866d7b-f153-4962-a751-c2744d06b165" containerID="904e528b2f302ad44050994e6c6aac295215a3d24aae9530ab47dee372a02927" exitCode=0 Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.830418 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" event={"ID":"aa866d7b-f153-4962-a751-c2744d06b165","Type":"ContainerDied","Data":"904e528b2f302ad44050994e6c6aac295215a3d24aae9530ab47dee372a02927"} Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.890093 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:31 crc kubenswrapper[4990]: I1203 13:46:31.937358 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fhv9c"] Dec 03 13:46:32 crc kubenswrapper[4990]: I1203 13:46:32.951718 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:46:32 crc kubenswrapper[4990]: I1203 13:46:32.984103 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-gjpjf"] Dec 03 13:46:32 crc kubenswrapper[4990]: I1203 13:46:32.992245 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-gjpjf"] Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.018643 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twwdv\" (UniqueName: \"kubernetes.io/projected/aa866d7b-f153-4962-a751-c2744d06b165-kube-api-access-twwdv\") pod \"aa866d7b-f153-4962-a751-c2744d06b165\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.018891 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa866d7b-f153-4962-a751-c2744d06b165-host\") pod \"aa866d7b-f153-4962-a751-c2744d06b165\" (UID: \"aa866d7b-f153-4962-a751-c2744d06b165\") " Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.018939 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa866d7b-f153-4962-a751-c2744d06b165-host" (OuterVolumeSpecName: "host") pod "aa866d7b-f153-4962-a751-c2744d06b165" (UID: "aa866d7b-f153-4962-a751-c2744d06b165"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.019332 4990 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa866d7b-f153-4962-a751-c2744d06b165-host\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.028606 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa866d7b-f153-4962-a751-c2744d06b165-kube-api-access-twwdv" (OuterVolumeSpecName: "kube-api-access-twwdv") pod "aa866d7b-f153-4962-a751-c2744d06b165" (UID: "aa866d7b-f153-4962-a751-c2744d06b165"). InnerVolumeSpecName "kube-api-access-twwdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.121690 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twwdv\" (UniqueName: \"kubernetes.io/projected/aa866d7b-f153-4962-a751-c2744d06b165-kube-api-access-twwdv\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.286118 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.286184 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.846643 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fhv9c" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="registry-server" containerID="cri-o://9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82" gracePeriod=2 Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.846929 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-gjpjf" Dec 03 13:46:33 crc kubenswrapper[4990]: I1203 13:46:33.848638 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc7a8ea362298dfcb481cbb564bc669a423026a454e2ebb8932cb304e279e39" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.244803 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-wjvd5"] Dec 03 13:46:34 crc kubenswrapper[4990]: E1203 13:46:34.245485 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa866d7b-f153-4962-a751-c2744d06b165" containerName="container-00" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.245499 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa866d7b-f153-4962-a751-c2744d06b165" containerName="container-00" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.245717 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa866d7b-f153-4962-a751-c2744d06b165" containerName="container-00" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.246530 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.278745 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa866d7b-f153-4962-a751-c2744d06b165" path="/var/lib/kubelet/pods/aa866d7b-f153-4962-a751-c2744d06b165/volumes" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.364230 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ac0ccbc-7d57-4574-823a-ec44432b01f0-host\") pod \"crc-debug-wjvd5\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.364290 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8br4\" (UniqueName: \"kubernetes.io/projected/9ac0ccbc-7d57-4574-823a-ec44432b01f0-kube-api-access-h8br4\") pod \"crc-debug-wjvd5\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.392102 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.465066 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kflhz\" (UniqueName: \"kubernetes.io/projected/d31b4fa7-9f88-490d-9b56-285739156f99-kube-api-access-kflhz\") pod \"d31b4fa7-9f88-490d-9b56-285739156f99\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.465322 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-catalog-content\") pod \"d31b4fa7-9f88-490d-9b56-285739156f99\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.465480 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-utilities\") pod \"d31b4fa7-9f88-490d-9b56-285739156f99\" (UID: \"d31b4fa7-9f88-490d-9b56-285739156f99\") " Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.465918 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ac0ccbc-7d57-4574-823a-ec44432b01f0-host\") pod \"crc-debug-wjvd5\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.465989 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ac0ccbc-7d57-4574-823a-ec44432b01f0-host\") pod \"crc-debug-wjvd5\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.466055 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-utilities" (OuterVolumeSpecName: "utilities") pod "d31b4fa7-9f88-490d-9b56-285739156f99" (UID: "d31b4fa7-9f88-490d-9b56-285739156f99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.466083 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8br4\" (UniqueName: \"kubernetes.io/projected/9ac0ccbc-7d57-4574-823a-ec44432b01f0-kube-api-access-h8br4\") pod \"crc-debug-wjvd5\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.466377 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.473070 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d31b4fa7-9f88-490d-9b56-285739156f99-kube-api-access-kflhz" (OuterVolumeSpecName: "kube-api-access-kflhz") pod "d31b4fa7-9f88-490d-9b56-285739156f99" (UID: "d31b4fa7-9f88-490d-9b56-285739156f99"). InnerVolumeSpecName "kube-api-access-kflhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.482241 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8br4\" (UniqueName: \"kubernetes.io/projected/9ac0ccbc-7d57-4574-823a-ec44432b01f0-kube-api-access-h8br4\") pod \"crc-debug-wjvd5\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.517470 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d31b4fa7-9f88-490d-9b56-285739156f99" (UID: "d31b4fa7-9f88-490d-9b56-285739156f99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.568531 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d31b4fa7-9f88-490d-9b56-285739156f99-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.568840 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kflhz\" (UniqueName: \"kubernetes.io/projected/d31b4fa7-9f88-490d-9b56-285739156f99-kube-api-access-kflhz\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.574344 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.856191 4990 generic.go:334] "Generic (PLEG): container finished" podID="9ac0ccbc-7d57-4574-823a-ec44432b01f0" containerID="bded836e533263c78a9abfbced3fa4af426d31211daef86cc4733636a4aef326" exitCode=0 Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.856288 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" event={"ID":"9ac0ccbc-7d57-4574-823a-ec44432b01f0","Type":"ContainerDied","Data":"bded836e533263c78a9abfbced3fa4af426d31211daef86cc4733636a4aef326"} Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.856482 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" event={"ID":"9ac0ccbc-7d57-4574-823a-ec44432b01f0","Type":"ContainerStarted","Data":"fe4c0285674cfcb81eaf3d41e0f1595b90e5b2f2ddf38f373a253b92d47a9526"} Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.859036 4990 generic.go:334] "Generic (PLEG): container finished" podID="d31b4fa7-9f88-490d-9b56-285739156f99" containerID="9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82" exitCode=0 Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.859081 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhv9c" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.859083 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhv9c" event={"ID":"d31b4fa7-9f88-490d-9b56-285739156f99","Type":"ContainerDied","Data":"9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82"} Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.859240 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhv9c" event={"ID":"d31b4fa7-9f88-490d-9b56-285739156f99","Type":"ContainerDied","Data":"eb996dfa971c20a022f6cde749035605747bad0d865d4ee7c483ad971d974688"} Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.859340 4990 scope.go:117] "RemoveContainer" containerID="9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.892061 4990 scope.go:117] "RemoveContainer" containerID="8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.907963 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fhv9c"] Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.915818 4990 scope.go:117] "RemoveContainer" containerID="e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.919787 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fhv9c"] Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.943705 4990 scope.go:117] "RemoveContainer" containerID="9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82" Dec 03 13:46:34 crc kubenswrapper[4990]: E1203 13:46:34.944268 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82\": container with ID starting with 9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82 not found: ID does not exist" containerID="9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.944299 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82"} err="failed to get container status \"9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82\": rpc error: code = NotFound desc = could not find container \"9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82\": container with ID starting with 9b0b2067f1be7dcdce42fe2677369796d4345b4d753111269725b9c1b0b1ae82 not found: ID does not exist" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.944320 4990 scope.go:117] "RemoveContainer" containerID="8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35" Dec 03 13:46:34 crc kubenswrapper[4990]: E1203 13:46:34.944615 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35\": container with ID starting with 8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35 not found: ID does not exist" containerID="8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.944639 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35"} err="failed to get container status \"8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35\": rpc error: code = NotFound desc = could not find container \"8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35\": container with ID starting with 8872c0531610f4f09778b374499230fdbcd07b9927e2bf8dceb5ea0f590fcf35 not found: ID does not exist" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.944653 4990 scope.go:117] "RemoveContainer" containerID="e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7" Dec 03 13:46:34 crc kubenswrapper[4990]: E1203 13:46:34.944881 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7\": container with ID starting with e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7 not found: ID does not exist" containerID="e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7" Dec 03 13:46:34 crc kubenswrapper[4990]: I1203 13:46:34.944941 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7"} err="failed to get container status \"e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7\": rpc error: code = NotFound desc = could not find container \"e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7\": container with ID starting with e15fdf636558dc23a34548b1e9f38202ebf80d99ee076c2843a57c042df4f1c7 not found: ID does not exist" Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.293859 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-wjvd5"] Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.301120 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-wjvd5"] Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.978751 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.995755 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8br4\" (UniqueName: \"kubernetes.io/projected/9ac0ccbc-7d57-4574-823a-ec44432b01f0-kube-api-access-h8br4\") pod \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.995913 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ac0ccbc-7d57-4574-823a-ec44432b01f0-host\") pod \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\" (UID: \"9ac0ccbc-7d57-4574-823a-ec44432b01f0\") " Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.996214 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ac0ccbc-7d57-4574-823a-ec44432b01f0-host" (OuterVolumeSpecName: "host") pod "9ac0ccbc-7d57-4574-823a-ec44432b01f0" (UID: "9ac0ccbc-7d57-4574-823a-ec44432b01f0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:46:35 crc kubenswrapper[4990]: I1203 13:46:35.996488 4990 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ac0ccbc-7d57-4574-823a-ec44432b01f0-host\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.000629 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac0ccbc-7d57-4574-823a-ec44432b01f0-kube-api-access-h8br4" (OuterVolumeSpecName: "kube-api-access-h8br4") pod "9ac0ccbc-7d57-4574-823a-ec44432b01f0" (UID: "9ac0ccbc-7d57-4574-823a-ec44432b01f0"). InnerVolumeSpecName "kube-api-access-h8br4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.098241 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8br4\" (UniqueName: \"kubernetes.io/projected/9ac0ccbc-7d57-4574-823a-ec44432b01f0-kube-api-access-h8br4\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.276986 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac0ccbc-7d57-4574-823a-ec44432b01f0" path="/var/lib/kubelet/pods/9ac0ccbc-7d57-4574-823a-ec44432b01f0/volumes" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.279923 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" path="/var/lib/kubelet/pods/d31b4fa7-9f88-490d-9b56-285739156f99/volumes" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.446212 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-54w97"] Dec 03 13:46:36 crc kubenswrapper[4990]: E1203 13:46:36.447672 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="extract-content" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.447784 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="extract-content" Dec 03 13:46:36 crc kubenswrapper[4990]: E1203 13:46:36.447899 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="registry-server" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.447958 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="registry-server" Dec 03 13:46:36 crc kubenswrapper[4990]: E1203 13:46:36.448037 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac0ccbc-7d57-4574-823a-ec44432b01f0" containerName="container-00" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.448131 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac0ccbc-7d57-4574-823a-ec44432b01f0" containerName="container-00" Dec 03 13:46:36 crc kubenswrapper[4990]: E1203 13:46:36.448202 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="extract-utilities" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.448260 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="extract-utilities" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.448513 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31b4fa7-9f88-490d-9b56-285739156f99" containerName="registry-server" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.448602 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac0ccbc-7d57-4574-823a-ec44432b01f0" containerName="container-00" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.449319 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.505414 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6caffd50-3991-4128-9a51-f8fc10b9b34b-host\") pod \"crc-debug-54w97\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.505593 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2vbb\" (UniqueName: \"kubernetes.io/projected/6caffd50-3991-4128-9a51-f8fc10b9b34b-kube-api-access-x2vbb\") pod \"crc-debug-54w97\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.607312 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2vbb\" (UniqueName: \"kubernetes.io/projected/6caffd50-3991-4128-9a51-f8fc10b9b34b-kube-api-access-x2vbb\") pod \"crc-debug-54w97\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.607494 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6caffd50-3991-4128-9a51-f8fc10b9b34b-host\") pod \"crc-debug-54w97\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.607632 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6caffd50-3991-4128-9a51-f8fc10b9b34b-host\") pod \"crc-debug-54w97\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.625890 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2vbb\" (UniqueName: \"kubernetes.io/projected/6caffd50-3991-4128-9a51-f8fc10b9b34b-kube-api-access-x2vbb\") pod \"crc-debug-54w97\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.766591 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:36 crc kubenswrapper[4990]: W1203 13:46:36.794665 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6caffd50_3991_4128_9a51_f8fc10b9b34b.slice/crio-ca38af8b53324afd766580f3b894a3d830eb9c775710bfce5eeb6512034e7874 WatchSource:0}: Error finding container ca38af8b53324afd766580f3b894a3d830eb9c775710bfce5eeb6512034e7874: Status 404 returned error can't find the container with id ca38af8b53324afd766580f3b894a3d830eb9c775710bfce5eeb6512034e7874 Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.876827 4990 scope.go:117] "RemoveContainer" containerID="bded836e533263c78a9abfbced3fa4af426d31211daef86cc4733636a4aef326" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.876947 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-wjvd5" Dec 03 13:46:36 crc kubenswrapper[4990]: I1203 13:46:36.880372 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-54w97" event={"ID":"6caffd50-3991-4128-9a51-f8fc10b9b34b","Type":"ContainerStarted","Data":"ca38af8b53324afd766580f3b894a3d830eb9c775710bfce5eeb6512034e7874"} Dec 03 13:46:37 crc kubenswrapper[4990]: I1203 13:46:37.891238 4990 generic.go:334] "Generic (PLEG): container finished" podID="6caffd50-3991-4128-9a51-f8fc10b9b34b" containerID="14f9dbcbad7ba46b0217768618e2be129377b5cda76b7382afe4aed7eb987b9c" exitCode=0 Dec 03 13:46:37 crc kubenswrapper[4990]: I1203 13:46:37.891397 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/crc-debug-54w97" event={"ID":"6caffd50-3991-4128-9a51-f8fc10b9b34b","Type":"ContainerDied","Data":"14f9dbcbad7ba46b0217768618e2be129377b5cda76b7382afe4aed7eb987b9c"} Dec 03 13:46:37 crc kubenswrapper[4990]: I1203 13:46:37.942740 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-54w97"] Dec 03 13:46:37 crc kubenswrapper[4990]: I1203 13:46:37.954541 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q8ss7/crc-debug-54w97"] Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.176560 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.358115 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6caffd50-3991-4128-9a51-f8fc10b9b34b-host\") pod \"6caffd50-3991-4128-9a51-f8fc10b9b34b\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.358250 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6caffd50-3991-4128-9a51-f8fc10b9b34b-host" (OuterVolumeSpecName: "host") pod "6caffd50-3991-4128-9a51-f8fc10b9b34b" (UID: "6caffd50-3991-4128-9a51-f8fc10b9b34b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.358381 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2vbb\" (UniqueName: \"kubernetes.io/projected/6caffd50-3991-4128-9a51-f8fc10b9b34b-kube-api-access-x2vbb\") pod \"6caffd50-3991-4128-9a51-f8fc10b9b34b\" (UID: \"6caffd50-3991-4128-9a51-f8fc10b9b34b\") " Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.359417 4990 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6caffd50-3991-4128-9a51-f8fc10b9b34b-host\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.363313 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6caffd50-3991-4128-9a51-f8fc10b9b34b-kube-api-access-x2vbb" (OuterVolumeSpecName: "kube-api-access-x2vbb") pod "6caffd50-3991-4128-9a51-f8fc10b9b34b" (UID: "6caffd50-3991-4128-9a51-f8fc10b9b34b"). InnerVolumeSpecName "kube-api-access-x2vbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.461298 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2vbb\" (UniqueName: \"kubernetes.io/projected/6caffd50-3991-4128-9a51-f8fc10b9b34b-kube-api-access-x2vbb\") on node \"crc\" DevicePath \"\"" Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.913400 4990 scope.go:117] "RemoveContainer" containerID="14f9dbcbad7ba46b0217768618e2be129377b5cda76b7382afe4aed7eb987b9c" Dec 03 13:46:39 crc kubenswrapper[4990]: I1203 13:46:39.913440 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/crc-debug-54w97" Dec 03 13:46:40 crc kubenswrapper[4990]: I1203 13:46:40.278795 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6caffd50-3991-4128-9a51-f8fc10b9b34b" path="/var/lib/kubelet/pods/6caffd50-3991-4128-9a51-f8fc10b9b34b/volumes" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.161847 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65cfc74cf4-n77zl_630b8058-02f8-4c7e-aeb7-6cca035356ed/barbican-api/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.222319 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65cfc74cf4-n77zl_630b8058-02f8-4c7e-aeb7-6cca035356ed/barbican-api-log/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.378249 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68db7d64bb-tmc56_fc68490b-9eed-4344-a09f-96adddd6c7c7/barbican-keystone-listener/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.390147 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68db7d64bb-tmc56_fc68490b-9eed-4344-a09f-96adddd6c7c7/barbican-keystone-listener-log/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.514262 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-67f9cc6bf5-jfjbk_8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b/barbican-worker/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.590627 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-67f9cc6bf5-jfjbk_8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b/barbican-worker-log/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.735519 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7_6c0707b9-5be8-4bf4-870b-ace30c9a9fe0/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.812827 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/ceilometer-central-agent/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.866470 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/ceilometer-notification-agent/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.937328 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/sg-core/0.log" Dec 03 13:46:55 crc kubenswrapper[4990]: I1203 13:46:55.960155 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/proxy-httpd/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.118793 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8534a0f8-e17f-49e3-948f-475e422d43e8/cinder-api/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.151849 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8534a0f8-e17f-49e3-948f-475e422d43e8/cinder-api-log/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.344772 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d724111e-b7f1-4efe-a05e-1e5cf2f5a278/cinder-scheduler/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.396080 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d724111e-b7f1-4efe-a05e-1e5cf2f5a278/probe/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.457308 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wltmb_7c2c7e25-8434-436c-8e35-61011e98873c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.668093 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm_8c311218-88d9-4022-8962-8ff68d03fa29/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.679945 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-s9vgm_f65f22f9-af79-4a63-bfc8-590d88507cb3/init/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.909051 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-s9vgm_f65f22f9-af79-4a63-bfc8-590d88507cb3/dnsmasq-dns/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.935955 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-s9vgm_f65f22f9-af79-4a63-bfc8-590d88507cb3/init/0.log" Dec 03 13:46:56 crc kubenswrapper[4990]: I1203 13:46:56.955347 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-57tqr_48109f08-9566-4bb0-bc9c-68413ddd32ae/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.144736 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fbdb7170-e955-4fd4-bfd5-03ad13363aa9/glance-httpd/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.204234 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fbdb7170-e955-4fd4-bfd5-03ad13363aa9/glance-log/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.360673 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_fd733a23-295d-4c67-9f67-1e27d99d7847/glance-httpd/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.408129 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_fd733a23-295d-4c67-9f67-1e27d99d7847/glance-log/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.491015 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5f6447d884-4db4z_c626101c-4745-40ce-aba1-587cb02ea499/horizon/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.743704 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v_213d788e-4e19-4020-a8ef-1d84d287fb95/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.936935 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5f6447d884-4db4z_c626101c-4745-40ce-aba1-587cb02ea499/horizon-log/0.log" Dec 03 13:46:57 crc kubenswrapper[4990]: I1203 13:46:57.938857 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-blhdl_a777faab-eccb-4a4f-b181-836edc76d745/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:58 crc kubenswrapper[4990]: I1203 13:46:58.181628 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6d44745d6d-9zbl5_6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383/keystone-api/0.log" Dec 03 13:46:58 crc kubenswrapper[4990]: I1203 13:46:58.298824 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b60c63f3-9d80-4660-9bc9-4afef4bd7cd6/kube-state-metrics/0.log" Dec 03 13:46:58 crc kubenswrapper[4990]: I1203 13:46:58.603091 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk_4c4068fd-3e0a-442a-b035-f2e84fba7e88/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:58 crc kubenswrapper[4990]: I1203 13:46:58.997130 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b4f6458f-phdfw_a486143f-bd72-4292-aaa1-194fb374862a/neutron-httpd/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.014109 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b4f6458f-phdfw_a486143f-bd72-4292-aaa1-194fb374862a/neutron-api/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.141662 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w_52371a09-a7eb-42bb-aa53-3a2188c6658d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.642847 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_11fb2cd9-0846-413e-b00a-29f258f641c5/nova-cell0-conductor-conductor/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.648534 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c704a76d-fc53-4741-9bd5-b7893e1c96a5/nova-api-log/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.901070 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c704a76d-fc53-4741-9bd5-b7893e1c96a5/nova-api-api/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.942544 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_a34660cc-b629-4ac1-ab68-0546243d600a/nova-cell1-conductor-conductor/0.log" Dec 03 13:46:59 crc kubenswrapper[4990]: I1203 13:46:59.963568 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_37a5c5c8-e837-456a-94ac-bd2d436b7e44/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 13:47:00 crc kubenswrapper[4990]: I1203 13:47:00.170774 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hhvts_4191ee3d-7722-4690-938c-7b8d8478589e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:00 crc kubenswrapper[4990]: I1203 13:47:00.336983 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f4737b59-9b3b-4bd5-8220-1346a05c998f/nova-metadata-log/0.log" Dec 03 13:47:00 crc kubenswrapper[4990]: I1203 13:47:00.624374 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_9e634575-20df-49bc-b2af-5322b408b702/nova-scheduler-scheduler/0.log" Dec 03 13:47:00 crc kubenswrapper[4990]: I1203 13:47:00.722686 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81e14867-9f93-4820-ba0f-c6dbd12e31db/mysql-bootstrap/0.log" Dec 03 13:47:00 crc kubenswrapper[4990]: I1203 13:47:00.888997 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81e14867-9f93-4820-ba0f-c6dbd12e31db/galera/0.log" Dec 03 13:47:00 crc kubenswrapper[4990]: I1203 13:47:00.952183 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81e14867-9f93-4820-ba0f-c6dbd12e31db/mysql-bootstrap/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.102428 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4fb6a200-beee-4b47-ac8d-e370e307070f/mysql-bootstrap/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.366913 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4fb6a200-beee-4b47-ac8d-e370e307070f/mysql-bootstrap/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.377279 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4fb6a200-beee-4b47-ac8d-e370e307070f/galera/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.607780 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f4737b59-9b3b-4bd5-8220-1346a05c998f/nova-metadata-metadata/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.612903 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9fdd0021-daac-4e15-b565-25cc720ed808/openstackclient/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.729879 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6frjn_83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5/ovn-controller/0.log" Dec 03 13:47:01 crc kubenswrapper[4990]: I1203 13:47:01.840327 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-dpmrl_caff07ff-f45e-4438-96a7-545d4de585c0/openstack-network-exporter/0.log" Dec 03 13:47:02 crc kubenswrapper[4990]: I1203 13:47:02.261294 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovsdb-server-init/0.log" Dec 03 13:47:02 crc kubenswrapper[4990]: I1203 13:47:02.338706 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovs-vswitchd/0.log" Dec 03 13:47:02 crc kubenswrapper[4990]: I1203 13:47:02.339098 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovsdb-server-init/0.log" Dec 03 13:47:02 crc kubenswrapper[4990]: I1203 13:47:02.380773 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovsdb-server/0.log" Dec 03 13:47:02 crc kubenswrapper[4990]: I1203 13:47:02.763809 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7d581ac6-ba40-4773-9d18-5aa2fcdfad62/openstack-network-exporter/0.log" Dec 03 13:47:02 crc kubenswrapper[4990]: I1203 13:47:02.870919 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-q8c6t_3cf5d887-0878-407b-98c9-539cb967a73e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.038918 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7d581ac6-ba40-4773-9d18-5aa2fcdfad62/ovn-northd/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.057016 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8ae85bd2-714b-4a62-b287-c2a9780753bd/openstack-network-exporter/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.188733 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8ae85bd2-714b-4a62-b287-c2a9780753bd/ovsdbserver-nb/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.267360 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_63ff1335-eac5-4524-b448-785488679bff/openstack-network-exporter/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.285969 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.286027 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.318651 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_63ff1335-eac5-4524-b448-785488679bff/ovsdbserver-sb/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.661671 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-644b9ffb4d-v4tbk_11b365c1-27dc-4c2f-a79c-12069d221f7f/placement-api/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.702428 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-644b9ffb4d-v4tbk_11b365c1-27dc-4c2f-a79c-12069d221f7f/placement-log/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.734366 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_be4855bd-7c56-4d1c-af0d-4017083409a6/setup-container/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.921803 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_be4855bd-7c56-4d1c-af0d-4017083409a6/setup-container/0.log" Dec 03 13:47:03 crc kubenswrapper[4990]: I1203 13:47:03.998672 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98e2d7f8-4691-44ca-a46f-9e08c82844a2/setup-container/0.log" Dec 03 13:47:04 crc kubenswrapper[4990]: I1203 13:47:04.004842 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_be4855bd-7c56-4d1c-af0d-4017083409a6/rabbitmq/0.log" Dec 03 13:47:04 crc kubenswrapper[4990]: I1203 13:47:04.775850 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98e2d7f8-4691-44ca-a46f-9e08c82844a2/rabbitmq/0.log" Dec 03 13:47:04 crc kubenswrapper[4990]: I1203 13:47:04.776314 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98e2d7f8-4691-44ca-a46f-9e08c82844a2/setup-container/0.log" Dec 03 13:47:04 crc kubenswrapper[4990]: I1203 13:47:04.827726 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p_5398edd4-615e-4e57-b25b-e7732ef5dbf3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:04 crc kubenswrapper[4990]: I1203 13:47:04.984841 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-d2cf7_748e28f7-7642-4591-bdba-29a79d86f5af/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:05 crc kubenswrapper[4990]: I1203 13:47:05.176277 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p972z_3c11a50b-b2d3-45be-9999-bcd93b698b1c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:05 crc kubenswrapper[4990]: I1203 13:47:05.462210 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-hbnrv_58dc37ce-d8bf-46f0-9338-43620480d0dd/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:05 crc kubenswrapper[4990]: I1203 13:47:05.632952 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ntk2h_dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf/ssh-known-hosts-edpm-deployment/0.log" Dec 03 13:47:05 crc kubenswrapper[4990]: I1203 13:47:05.858354 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86d5555b99-xbxkx_62fcb6e6-b158-4a76-8570-3563346e2a77/proxy-server/0.log" Dec 03 13:47:05 crc kubenswrapper[4990]: I1203 13:47:05.892597 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86d5555b99-xbxkx_62fcb6e6-b158-4a76-8570-3563346e2a77/proxy-httpd/0.log" Dec 03 13:47:05 crc kubenswrapper[4990]: I1203 13:47:05.931348 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gqdz2_73f3c670-ab5f-4ab9-9278-9aa008c9a35b/swift-ring-rebalance/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.094133 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-auditor/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.164431 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-reaper/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.213990 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-replicator/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.338855 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-server/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.351284 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-auditor/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.420006 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-replicator/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.503197 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-server/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.584367 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-updater/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.610873 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-auditor/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.709244 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-expirer/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.725938 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-replicator/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.793126 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-server/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.831920 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-updater/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.928757 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/rsync/0.log" Dec 03 13:47:06 crc kubenswrapper[4990]: I1203 13:47:06.992238 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/swift-recon-cron/0.log" Dec 03 13:47:07 crc kubenswrapper[4990]: I1203 13:47:07.130629 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9_882b0d7e-4fd5-4462-ba56-091dbc993c0b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:07 crc kubenswrapper[4990]: I1203 13:47:07.316058 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_f375426e-6095-4621-ab29-6c7324da86e7/tempest-tests-tempest-tests-runner/0.log" Dec 03 13:47:07 crc kubenswrapper[4990]: I1203 13:47:07.358919 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f0fab240-b45f-44e1-8fb7-446c4dc6d64e/test-operator-logs-container/0.log" Dec 03 13:47:07 crc kubenswrapper[4990]: I1203 13:47:07.592950 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv_5207255c-c5b2-470d-aa44-57da1579fe01/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:47:16 crc kubenswrapper[4990]: I1203 13:47:16.648706 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_24350652-92e4-40b6-82d5-4b785a2a0314/memcached/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.266708 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4sp89_c7a7a3b2-d356-4dfa-b640-ae39aa2d4840/kube-rbac-proxy/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.285880 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.285933 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.285970 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.286630 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33931fd5dcd79ed868236bde132bdc93fe9fd197d027f2241b26f817e20a5e50"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.286676 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://33931fd5dcd79ed868236bde132bdc93fe9fd197d027f2241b26f817e20a5e50" gracePeriod=600 Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.322188 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4sp89_c7a7a3b2-d356-4dfa-b640-ae39aa2d4840/manager/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.435289 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="33931fd5dcd79ed868236bde132bdc93fe9fd197d027f2241b26f817e20a5e50" exitCode=0 Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.435349 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"33931fd5dcd79ed868236bde132bdc93fe9fd197d027f2241b26f817e20a5e50"} Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.435396 4990 scope.go:117] "RemoveContainer" containerID="f4c45b6dc83e53502aa9898f78353468ca2042969b22bf9bf324778f2c166173" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.469437 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7n65k_c45d3b36-d188-4bb1-b19f-5521709f572a/kube-rbac-proxy/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.570646 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7n65k_c45d3b36-d188-4bb1-b19f-5521709f572a/manager/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.665020 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hqhfc_93f0e562-5973-4c2a-ab48-aa167ae49ffa/kube-rbac-proxy/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.669176 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hqhfc_93f0e562-5973-4c2a-ab48-aa167ae49ffa/manager/0.log" Dec 03 13:47:33 crc kubenswrapper[4990]: I1203 13:47:33.765907 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/util/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.003857 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/util/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.026075 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/pull/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.052936 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/pull/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.233405 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/util/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.244914 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/extract/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.275703 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/pull/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.428165 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-z484w_ac715eff-4f7f-44b0-b5ca-c5d283edca36/kube-rbac-proxy/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.447590 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f"} Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.532942 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-z484w_ac715eff-4f7f-44b0-b5ca-c5d283edca36/manager/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.538184 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xt2lq_a8fb4bb3-4b67-4ada-8876-389bff0e9836/kube-rbac-proxy/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.661324 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xt2lq_a8fb4bb3-4b67-4ada-8876-389bff0e9836/manager/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.740768 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-8zwv5_4881a1ed-a968-4f20-a8e6-94a91ec5eceb/kube-rbac-proxy/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.745419 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-8zwv5_4881a1ed-a968-4f20-a8e6-94a91ec5eceb/manager/0.log" Dec 03 13:47:34 crc kubenswrapper[4990]: I1203 13:47:34.945748 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zhgnf_0e880282-b396-4fa9-a3ac-89d81315ecd8/kube-rbac-proxy/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.121738 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zhgnf_0e880282-b396-4fa9-a3ac-89d81315ecd8/manager/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.121076 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-79xxb_96da719f-5659-45eb-ac48-37165c936527/kube-rbac-proxy/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.145462 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-79xxb_96da719f-5659-45eb-ac48-37165c936527/manager/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.372545 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-256rj_43c2b7af-9659-4ae6-8809-27475b31c611/kube-rbac-proxy/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.381192 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-256rj_43c2b7af-9659-4ae6-8809-27475b31c611/manager/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.440657 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qt45w_9106e41f-5788-46a2-86ce-6362832d6df7/kube-rbac-proxy/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.582132 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qt45w_9106e41f-5788-46a2-86ce-6362832d6df7/manager/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.641084 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vkrn9_e944cb60-4e1a-448f-ab25-60ff87e3a166/kube-rbac-proxy/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.706718 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vkrn9_e944cb60-4e1a-448f-ab25-60ff87e3a166/manager/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.824666 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-c8cgl_0a2be3dd-196c-4186-ac5d-85ce550aa201/kube-rbac-proxy/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.894127 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-c8cgl_0a2be3dd-196c-4186-ac5d-85ce550aa201/manager/0.log" Dec 03 13:47:35 crc kubenswrapper[4990]: I1203 13:47:35.995285 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dg8m2_a18809dc-c194-44b4-84e8-26e6da515bb7/kube-rbac-proxy/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.117523 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dg8m2_a18809dc-c194-44b4-84e8-26e6da515bb7/manager/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.170851 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-xg94h_6ca15749-37c2-477f-8a9a-33c07b97aaf3/kube-rbac-proxy/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.277580 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-xg94h_6ca15749-37c2-477f-8a9a-33c07b97aaf3/manager/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.382212 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9_c0833459-1161-4beb-ad68-07d51b5b33d7/kube-rbac-proxy/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.394773 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9_c0833459-1161-4beb-ad68-07d51b5b33d7/manager/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.803418 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85bd6bf7bc-zdj92_49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d/operator/0.log" Dec 03 13:47:36 crc kubenswrapper[4990]: I1203 13:47:36.808523 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xgxsw_c1b91baf-8945-490b-987c-531d89259d3b/registry-server/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.041119 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-4db2x_64f072f3-940c-431d-8e0f-b77e9349e79e/kube-rbac-proxy/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.063068 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-4db2x_64f072f3-940c-431d-8e0f-b77e9349e79e/manager/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.251222 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-x7jtx_3443975f-d482-422a-a010-9940c2b1182f/kube-rbac-proxy/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.317192 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-p8hpt_cc2d7750-b3b5-4585-95fd-4f84e1c437ad/operator/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.328718 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-x7jtx_3443975f-d482-422a-a010-9940c2b1182f/manager/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.509204 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-rvslq_8bf181cd-d802-43c1-a7f0-7df5ec5008a1/kube-rbac-proxy/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.605378 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-rvslq_8bf181cd-d802-43c1-a7f0-7df5ec5008a1/manager/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.749443 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5b474d9948-p5wjj_425132be-e04e-46f6-ac06-1546c5b12a26/manager/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.766761 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-4wkxw_213fe9a3-0270-4ec5-b380-2f28159b5b6a/kube-rbac-proxy/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.807129 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-4wkxw_213fe9a3-0270-4ec5-b380-2f28159b5b6a/manager/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.910342 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tg4ww_4481645b-1e70-415a-a01b-131b1761dd19/kube-rbac-proxy/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.952724 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tg4ww_4481645b-1e70-415a-a01b-131b1761dd19/manager/0.log" Dec 03 13:47:37 crc kubenswrapper[4990]: I1203 13:47:37.990566 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7cl8r_d2c9d382-f60c-4c2c-a684-ddec4371b165/kube-rbac-proxy/0.log" Dec 03 13:47:38 crc kubenswrapper[4990]: I1203 13:47:38.066858 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7cl8r_d2c9d382-f60c-4c2c-a684-ddec4371b165/manager/0.log" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.234228 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d7vnt"] Dec 03 13:47:51 crc kubenswrapper[4990]: E1203 13:47:51.235260 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6caffd50-3991-4128-9a51-f8fc10b9b34b" containerName="container-00" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.235277 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="6caffd50-3991-4128-9a51-f8fc10b9b34b" containerName="container-00" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.235535 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="6caffd50-3991-4128-9a51-f8fc10b9b34b" containerName="container-00" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.237234 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.249652 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d7vnt"] Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.353044 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-catalog-content\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.353279 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-utilities\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.353633 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm4xl\" (UniqueName: \"kubernetes.io/projected/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-kube-api-access-fm4xl\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.455628 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-catalog-content\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.455747 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-utilities\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.455826 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm4xl\" (UniqueName: \"kubernetes.io/projected/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-kube-api-access-fm4xl\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.456252 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-catalog-content\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.456344 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-utilities\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.476404 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm4xl\" (UniqueName: \"kubernetes.io/projected/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-kube-api-access-fm4xl\") pod \"redhat-operators-d7vnt\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.558113 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.844677 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2n5z2"] Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.847727 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.856841 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2n5z2"] Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.970825 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2xmv\" (UniqueName: \"kubernetes.io/projected/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-kube-api-access-j2xmv\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.971016 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-utilities\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:51 crc kubenswrapper[4990]: I1203 13:47:51.971086 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-catalog-content\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: W1203 13:47:52.056896 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc074cc5_41a4_41f4_bde6_d51f7eb5a1a4.slice/crio-a128c6a6b8ab7f0747b79636130e18abfff866ca505f08d7e36713a45c892829 WatchSource:0}: Error finding container a128c6a6b8ab7f0747b79636130e18abfff866ca505f08d7e36713a45c892829: Status 404 returned error can't find the container with id a128c6a6b8ab7f0747b79636130e18abfff866ca505f08d7e36713a45c892829 Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.072893 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-catalog-content\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.073005 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2xmv\" (UniqueName: \"kubernetes.io/projected/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-kube-api-access-j2xmv\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.073148 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-utilities\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.073849 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-utilities\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.074153 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-catalog-content\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.075347 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d7vnt"] Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.098809 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2xmv\" (UniqueName: \"kubernetes.io/projected/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-kube-api-access-j2xmv\") pod \"certified-operators-2n5z2\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.180525 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.624593 4990 generic.go:334] "Generic (PLEG): container finished" podID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerID="9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d" exitCode=0 Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.624829 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d7vnt" event={"ID":"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4","Type":"ContainerDied","Data":"9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d"} Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.624927 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d7vnt" event={"ID":"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4","Type":"ContainerStarted","Data":"a128c6a6b8ab7f0747b79636130e18abfff866ca505f08d7e36713a45c892829"} Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.626653 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:47:52 crc kubenswrapper[4990]: I1203 13:47:52.718824 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2n5z2"] Dec 03 13:47:53 crc kubenswrapper[4990]: I1203 13:47:53.636503 4990 generic.go:334] "Generic (PLEG): container finished" podID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerID="7cb0c5af181dcddc046555a8b15fb618f34269737c835b6f6f6562c2cf2aeca2" exitCode=0 Dec 03 13:47:53 crc kubenswrapper[4990]: I1203 13:47:53.636610 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerDied","Data":"7cb0c5af181dcddc046555a8b15fb618f34269737c835b6f6f6562c2cf2aeca2"} Dec 03 13:47:53 crc kubenswrapper[4990]: I1203 13:47:53.637070 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerStarted","Data":"ac76f131672e2c8b1edef09f4c6b5a4c77f8769b97cf6d5a9bd6e1a160c4392d"} Dec 03 13:47:54 crc kubenswrapper[4990]: I1203 13:47:54.648042 4990 generic.go:334] "Generic (PLEG): container finished" podID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerID="c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde" exitCode=0 Dec 03 13:47:54 crc kubenswrapper[4990]: I1203 13:47:54.648131 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d7vnt" event={"ID":"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4","Type":"ContainerDied","Data":"c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde"} Dec 03 13:47:54 crc kubenswrapper[4990]: I1203 13:47:54.657248 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerStarted","Data":"349ae328ad44964724d6168728cc9204bf6c906e6b9029bc2e697f44f8bd56f7"} Dec 03 13:47:55 crc kubenswrapper[4990]: I1203 13:47:55.666382 4990 generic.go:334] "Generic (PLEG): container finished" podID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerID="349ae328ad44964724d6168728cc9204bf6c906e6b9029bc2e697f44f8bd56f7" exitCode=0 Dec 03 13:47:55 crc kubenswrapper[4990]: I1203 13:47:55.666486 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerDied","Data":"349ae328ad44964724d6168728cc9204bf6c906e6b9029bc2e697f44f8bd56f7"} Dec 03 13:47:57 crc kubenswrapper[4990]: I1203 13:47:57.687773 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d7vnt" event={"ID":"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4","Type":"ContainerStarted","Data":"fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8"} Dec 03 13:47:58 crc kubenswrapper[4990]: I1203 13:47:58.698865 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jsjk4_a36bdb31-4352-4ef2-9d56-85cd07b89e21/control-plane-machine-set-operator/0.log" Dec 03 13:47:58 crc kubenswrapper[4990]: I1203 13:47:58.939748 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jfgrs_184ef2da-fe68-4fbc-8a55-f6eb63262764/kube-rbac-proxy/0.log" Dec 03 13:47:58 crc kubenswrapper[4990]: I1203 13:47:58.941359 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jfgrs_184ef2da-fe68-4fbc-8a55-f6eb63262764/machine-api-operator/0.log" Dec 03 13:47:59 crc kubenswrapper[4990]: I1203 13:47:59.709579 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerStarted","Data":"019469bfe6398455cb9746978244a4b8d8dd4f942fe848089c5aa69cf8f79544"} Dec 03 13:47:59 crc kubenswrapper[4990]: I1203 13:47:59.733367 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d7vnt" podStartSLOduration=5.388728183 podStartE2EDuration="8.733347058s" podCreationTimestamp="2025-12-03 13:47:51 +0000 UTC" firstStartedPulling="2025-12-03 13:47:52.626321458 +0000 UTC m=+4220.768232687" lastFinishedPulling="2025-12-03 13:47:55.970940313 +0000 UTC m=+4224.112851562" observedRunningTime="2025-12-03 13:47:57.713702333 +0000 UTC m=+4225.855613562" watchObservedRunningTime="2025-12-03 13:47:59.733347058 +0000 UTC m=+4227.875258287" Dec 03 13:47:59 crc kubenswrapper[4990]: I1203 13:47:59.739525 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2n5z2" podStartSLOduration=4.222680359 podStartE2EDuration="8.739505283s" podCreationTimestamp="2025-12-03 13:47:51 +0000 UTC" firstStartedPulling="2025-12-03 13:47:53.638819672 +0000 UTC m=+4221.780730901" lastFinishedPulling="2025-12-03 13:47:58.155644596 +0000 UTC m=+4226.297555825" observedRunningTime="2025-12-03 13:47:59.729551027 +0000 UTC m=+4227.871462256" watchObservedRunningTime="2025-12-03 13:47:59.739505283 +0000 UTC m=+4227.881416502" Dec 03 13:48:01 crc kubenswrapper[4990]: I1203 13:48:01.558551 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:48:01 crc kubenswrapper[4990]: I1203 13:48:01.559128 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:48:02 crc kubenswrapper[4990]: I1203 13:48:02.180922 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:48:02 crc kubenswrapper[4990]: I1203 13:48:02.180992 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:48:02 crc kubenswrapper[4990]: I1203 13:48:02.226826 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:48:02 crc kubenswrapper[4990]: I1203 13:48:02.614809 4990 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d7vnt" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="registry-server" probeResult="failure" output=< Dec 03 13:48:02 crc kubenswrapper[4990]: timeout: failed to connect service ":50051" within 1s Dec 03 13:48:02 crc kubenswrapper[4990]: > Dec 03 13:48:11 crc kubenswrapper[4990]: I1203 13:48:11.610095 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:48:11 crc kubenswrapper[4990]: I1203 13:48:11.659022 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:48:11 crc kubenswrapper[4990]: I1203 13:48:11.852548 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d7vnt"] Dec 03 13:48:12 crc kubenswrapper[4990]: I1203 13:48:12.239625 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:48:12 crc kubenswrapper[4990]: I1203 13:48:12.354582 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xrqmv_555f5cb7-9649-4c3a-8b77-244edcb83c7f/cert-manager-controller/0.log" Dec 03 13:48:12 crc kubenswrapper[4990]: I1203 13:48:12.542868 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-xlfdw_9dd6b4f0-81be-480c-84d6-22ba5b462ab0/cert-manager-cainjector/0.log" Dec 03 13:48:12 crc kubenswrapper[4990]: I1203 13:48:12.575287 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8ghd6_3196eb4c-4adf-493f-ade6-a9c839d0e4e6/cert-manager-webhook/0.log" Dec 03 13:48:12 crc kubenswrapper[4990]: I1203 13:48:12.829217 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d7vnt" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="registry-server" containerID="cri-o://fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8" gracePeriod=2 Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.311407 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.487526 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm4xl\" (UniqueName: \"kubernetes.io/projected/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-kube-api-access-fm4xl\") pod \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.487684 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-catalog-content\") pod \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.487723 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-utilities\") pod \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\" (UID: \"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4\") " Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.488787 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-utilities" (OuterVolumeSpecName: "utilities") pod "fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" (UID: "fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.507284 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-kube-api-access-fm4xl" (OuterVolumeSpecName: "kube-api-access-fm4xl") pod "fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" (UID: "fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4"). InnerVolumeSpecName "kube-api-access-fm4xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.593322 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.593363 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm4xl\" (UniqueName: \"kubernetes.io/projected/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-kube-api-access-fm4xl\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.618589 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" (UID: "fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.695738 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.841425 4990 generic.go:334] "Generic (PLEG): container finished" podID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerID="fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8" exitCode=0 Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.841486 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d7vnt" event={"ID":"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4","Type":"ContainerDied","Data":"fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8"} Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.841519 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d7vnt" event={"ID":"fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4","Type":"ContainerDied","Data":"a128c6a6b8ab7f0747b79636130e18abfff866ca505f08d7e36713a45c892829"} Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.841539 4990 scope.go:117] "RemoveContainer" containerID="fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.841702 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d7vnt" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.865668 4990 scope.go:117] "RemoveContainer" containerID="c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.892563 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d7vnt"] Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.903152 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d7vnt"] Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.916917 4990 scope.go:117] "RemoveContainer" containerID="9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.954663 4990 scope.go:117] "RemoveContainer" containerID="fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8" Dec 03 13:48:13 crc kubenswrapper[4990]: E1203 13:48:13.955933 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8\": container with ID starting with fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8 not found: ID does not exist" containerID="fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.955981 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8"} err="failed to get container status \"fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8\": rpc error: code = NotFound desc = could not find container \"fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8\": container with ID starting with fe3798eee3c746b2baad1d0db42e6dccd1bf10d4f5e8bd45954d8c438e3b42e8 not found: ID does not exist" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.956012 4990 scope.go:117] "RemoveContainer" containerID="c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde" Dec 03 13:48:13 crc kubenswrapper[4990]: E1203 13:48:13.957067 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde\": container with ID starting with c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde not found: ID does not exist" containerID="c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.957220 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde"} err="failed to get container status \"c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde\": rpc error: code = NotFound desc = could not find container \"c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde\": container with ID starting with c4a76353ccc2bf25092805490a7307056ecef1503b699c525817218a7544acde not found: ID does not exist" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.957292 4990 scope.go:117] "RemoveContainer" containerID="9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d" Dec 03 13:48:13 crc kubenswrapper[4990]: E1203 13:48:13.957889 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d\": container with ID starting with 9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d not found: ID does not exist" containerID="9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d" Dec 03 13:48:13 crc kubenswrapper[4990]: I1203 13:48:13.957949 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d"} err="failed to get container status \"9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d\": rpc error: code = NotFound desc = could not find container \"9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d\": container with ID starting with 9621e918cb83afc58ad4c7fcf4fd7dad7316523aa40264912645705f325ba53d not found: ID does not exist" Dec 03 13:48:14 crc kubenswrapper[4990]: I1203 13:48:14.276531 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" path="/var/lib/kubelet/pods/fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4/volumes" Dec 03 13:48:14 crc kubenswrapper[4990]: I1203 13:48:14.664752 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2n5z2"] Dec 03 13:48:14 crc kubenswrapper[4990]: I1203 13:48:14.665133 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2n5z2" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="registry-server" containerID="cri-o://019469bfe6398455cb9746978244a4b8d8dd4f942fe848089c5aa69cf8f79544" gracePeriod=2 Dec 03 13:48:14 crc kubenswrapper[4990]: I1203 13:48:14.854932 4990 generic.go:334] "Generic (PLEG): container finished" podID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerID="019469bfe6398455cb9746978244a4b8d8dd4f942fe848089c5aa69cf8f79544" exitCode=0 Dec 03 13:48:14 crc kubenswrapper[4990]: I1203 13:48:14.855011 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerDied","Data":"019469bfe6398455cb9746978244a4b8d8dd4f942fe848089c5aa69cf8f79544"} Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.163710 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.235590 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-catalog-content\") pod \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.235656 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-utilities\") pod \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.235774 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2xmv\" (UniqueName: \"kubernetes.io/projected/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-kube-api-access-j2xmv\") pod \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\" (UID: \"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0\") " Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.236482 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-utilities" (OuterVolumeSpecName: "utilities") pod "21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" (UID: "21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.239906 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.243758 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-kube-api-access-j2xmv" (OuterVolumeSpecName: "kube-api-access-j2xmv") pod "21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" (UID: "21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0"). InnerVolumeSpecName "kube-api-access-j2xmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.280793 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" (UID: "21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.341782 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.341822 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2xmv\" (UniqueName: \"kubernetes.io/projected/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0-kube-api-access-j2xmv\") on node \"crc\" DevicePath \"\"" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.867786 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2n5z2" event={"ID":"21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0","Type":"ContainerDied","Data":"ac76f131672e2c8b1edef09f4c6b5a4c77f8769b97cf6d5a9bd6e1a160c4392d"} Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.867846 4990 scope.go:117] "RemoveContainer" containerID="019469bfe6398455cb9746978244a4b8d8dd4f942fe848089c5aa69cf8f79544" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.867924 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2n5z2" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.891395 4990 scope.go:117] "RemoveContainer" containerID="349ae328ad44964724d6168728cc9204bf6c906e6b9029bc2e697f44f8bd56f7" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.909672 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2n5z2"] Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.910129 4990 scope.go:117] "RemoveContainer" containerID="7cb0c5af181dcddc046555a8b15fb618f34269737c835b6f6f6562c2cf2aeca2" Dec 03 13:48:15 crc kubenswrapper[4990]: I1203 13:48:15.919424 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2n5z2"] Dec 03 13:48:16 crc kubenswrapper[4990]: I1203 13:48:16.275630 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" path="/var/lib/kubelet/pods/21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0/volumes" Dec 03 13:48:24 crc kubenswrapper[4990]: I1203 13:48:24.608268 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7n2ph_c717132b-5c00-4d42-9d18-c163e82b27ff/nmstate-console-plugin/0.log" Dec 03 13:48:24 crc kubenswrapper[4990]: I1203 13:48:24.798937 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-gwfbg_afaa9abc-cdbc-44b9-ab3e-a69df533460c/kube-rbac-proxy/0.log" Dec 03 13:48:24 crc kubenswrapper[4990]: I1203 13:48:24.800559 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dtt26_20a68e4a-3638-430c-a27b-7d535613c3d3/nmstate-handler/0.log" Dec 03 13:48:24 crc kubenswrapper[4990]: I1203 13:48:24.870618 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-gwfbg_afaa9abc-cdbc-44b9-ab3e-a69df533460c/nmstate-metrics/0.log" Dec 03 13:48:24 crc kubenswrapper[4990]: I1203 13:48:24.984808 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-g69mt_9429610d-e1cb-4248-8224-12e08acf1a69/nmstate-operator/0.log" Dec 03 13:48:25 crc kubenswrapper[4990]: I1203 13:48:25.055779 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-pgjpk_45d5180b-e7cb-481e-88c7-9fa11bf8edc9/nmstate-webhook/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.479189 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hhxt8_9d94113a-0245-4057-97e6-8464ef2a830d/kube-rbac-proxy/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.632627 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hhxt8_9d94113a-0245-4057-97e6-8464ef2a830d/controller/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.691695 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.846907 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.879436 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.896267 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:48:40 crc kubenswrapper[4990]: I1203 13:48:40.926101 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.084439 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.126178 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.167434 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.188551 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.327938 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.333307 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.345570 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.376557 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/controller/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.526740 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/frr-metrics/0.log" Dec 03 13:48:41 crc kubenswrapper[4990]: I1203 13:48:41.529930 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/kube-rbac-proxy/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.265128 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/kube-rbac-proxy-frr/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.265853 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/reloader/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.469851 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-xnhk6_d64e42b3-c578-44d9-818b-3b0aa7f4b3e4/frr-k8s-webhook-server/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.594685 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-584cf47849-n84tl_5aac7d67-caa8-481d-80fb-7ca848f1007f/manager/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.750836 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-8464f887b-s5zch_1d95a171-2b09-4441-b7e7-31a5f4427a50/webhook-server/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.878263 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/frr/0.log" Dec 03 13:48:42 crc kubenswrapper[4990]: I1203 13:48:42.928429 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-t6cl8_10554cf0-1dc9-4e97-973f-1f2d47e3d1f8/kube-rbac-proxy/0.log" Dec 03 13:48:43 crc kubenswrapper[4990]: I1203 13:48:43.262381 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-t6cl8_10554cf0-1dc9-4e97-973f-1f2d47e3d1f8/speaker/0.log" Dec 03 13:48:54 crc kubenswrapper[4990]: I1203 13:48:54.973465 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/util/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.133385 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/pull/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.144569 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/util/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.162952 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/pull/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.328848 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/pull/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.344296 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/extract/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.374148 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/util/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.526764 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/util/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.686628 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/util/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.755405 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/pull/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.773518 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/pull/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.855499 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/pull/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.894642 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/util/0.log" Dec 03 13:48:55 crc kubenswrapper[4990]: I1203 13:48:55.982610 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/extract/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.021236 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-utilities/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.225161 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-content/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.233111 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-utilities/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.259904 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-content/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.386592 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-utilities/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.418219 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-content/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.638551 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-utilities/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.793124 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-utilities/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.851139 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-content/0.log" Dec 03 13:48:56 crc kubenswrapper[4990]: I1203 13:48:56.869195 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-content/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.048789 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/registry-server/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.065925 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-content/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.105465 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-utilities/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.290148 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9n6hl_8a0b82ca-6250-4ca2-abb5-2704907e4f98/marketplace-operator/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.560008 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-utilities/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.778170 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-content/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.788018 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-utilities/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.828013 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-content/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.854981 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/registry-server/0.log" Dec 03 13:48:57 crc kubenswrapper[4990]: I1203 13:48:57.985186 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-utilities/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.012719 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-content/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.197958 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/registry-server/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.227050 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-utilities/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.400179 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-content/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.413309 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-content/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.431687 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-utilities/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.590172 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-content/0.log" Dec 03 13:48:58 crc kubenswrapper[4990]: I1203 13:48:58.623728 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-utilities/0.log" Dec 03 13:48:59 crc kubenswrapper[4990]: I1203 13:48:59.235902 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/registry-server/0.log" Dec 03 13:49:27 crc kubenswrapper[4990]: I1203 13:49:27.171175 4990 scope.go:117] "RemoveContainer" containerID="04badae7c8d392d6974b50394c9b509aef3eb6a9aa668f3b5baaa241cd872500" Dec 03 13:49:27 crc kubenswrapper[4990]: I1203 13:49:27.216369 4990 scope.go:117] "RemoveContainer" containerID="ee5bd0dd67b87caa22aedbf58cfc5256e843e3f210c03bfb08260b720f532c5f" Dec 03 13:49:27 crc kubenswrapper[4990]: I1203 13:49:27.244133 4990 scope.go:117] "RemoveContainer" containerID="cfed5ebbd47cc23cb6e64698a533adaba50880d9ca8b60efe9770541dc4e3e0d" Dec 03 13:50:03 crc kubenswrapper[4990]: I1203 13:50:03.286282 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:50:03 crc kubenswrapper[4990]: I1203 13:50:03.288414 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:50:33 crc kubenswrapper[4990]: I1203 13:50:33.285951 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:50:33 crc kubenswrapper[4990]: I1203 13:50:33.286517 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:50:44 crc kubenswrapper[4990]: I1203 13:50:44.322179 4990 generic.go:334] "Generic (PLEG): container finished" podID="741c2504-9580-4514-bf3a-0de631f99c03" containerID="ab6b98faaa61f3d31fd4cd7cc909fbeb310c5a75efdeb6291f6a98ee258851ba" exitCode=0 Dec 03 13:50:44 crc kubenswrapper[4990]: I1203 13:50:44.322367 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q8ss7/must-gather-fd98h" event={"ID":"741c2504-9580-4514-bf3a-0de631f99c03","Type":"ContainerDied","Data":"ab6b98faaa61f3d31fd4cd7cc909fbeb310c5a75efdeb6291f6a98ee258851ba"} Dec 03 13:50:44 crc kubenswrapper[4990]: I1203 13:50:44.323355 4990 scope.go:117] "RemoveContainer" containerID="ab6b98faaa61f3d31fd4cd7cc909fbeb310c5a75efdeb6291f6a98ee258851ba" Dec 03 13:50:45 crc kubenswrapper[4990]: I1203 13:50:45.375595 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q8ss7_must-gather-fd98h_741c2504-9580-4514-bf3a-0de631f99c03/gather/0.log" Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.222863 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q8ss7/must-gather-fd98h"] Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.223674 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-q8ss7/must-gather-fd98h" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="copy" containerID="cri-o://51ca4fa7d11dcbc0f1159614a31c6413d6a421f99631397e2dcc509b856fa33e" gracePeriod=2 Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.233699 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q8ss7/must-gather-fd98h"] Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.420035 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q8ss7_must-gather-fd98h_741c2504-9580-4514-bf3a-0de631f99c03/copy/0.log" Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.420905 4990 generic.go:334] "Generic (PLEG): container finished" podID="741c2504-9580-4514-bf3a-0de631f99c03" containerID="51ca4fa7d11dcbc0f1159614a31c6413d6a421f99631397e2dcc509b856fa33e" exitCode=143 Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.805277 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q8ss7_must-gather-fd98h_741c2504-9580-4514-bf3a-0de631f99c03/copy/0.log" Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.805917 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.888661 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/741c2504-9580-4514-bf3a-0de631f99c03-must-gather-output\") pod \"741c2504-9580-4514-bf3a-0de631f99c03\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.888744 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mj8t\" (UniqueName: \"kubernetes.io/projected/741c2504-9580-4514-bf3a-0de631f99c03-kube-api-access-4mj8t\") pod \"741c2504-9580-4514-bf3a-0de631f99c03\" (UID: \"741c2504-9580-4514-bf3a-0de631f99c03\") " Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.901854 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/741c2504-9580-4514-bf3a-0de631f99c03-kube-api-access-4mj8t" (OuterVolumeSpecName: "kube-api-access-4mj8t") pod "741c2504-9580-4514-bf3a-0de631f99c03" (UID: "741c2504-9580-4514-bf3a-0de631f99c03"). InnerVolumeSpecName "kube-api-access-4mj8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:50:53 crc kubenswrapper[4990]: I1203 13:50:53.990891 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mj8t\" (UniqueName: \"kubernetes.io/projected/741c2504-9580-4514-bf3a-0de631f99c03-kube-api-access-4mj8t\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.032634 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/741c2504-9580-4514-bf3a-0de631f99c03-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "741c2504-9580-4514-bf3a-0de631f99c03" (UID: "741c2504-9580-4514-bf3a-0de631f99c03"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.092525 4990 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/741c2504-9580-4514-bf3a-0de631f99c03-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.276350 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="741c2504-9580-4514-bf3a-0de631f99c03" path="/var/lib/kubelet/pods/741c2504-9580-4514-bf3a-0de631f99c03/volumes" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.430864 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q8ss7_must-gather-fd98h_741c2504-9580-4514-bf3a-0de631f99c03/copy/0.log" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.431635 4990 scope.go:117] "RemoveContainer" containerID="51ca4fa7d11dcbc0f1159614a31c6413d6a421f99631397e2dcc509b856fa33e" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.431904 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q8ss7/must-gather-fd98h" Dec 03 13:50:54 crc kubenswrapper[4990]: I1203 13:50:54.516647 4990 scope.go:117] "RemoveContainer" containerID="ab6b98faaa61f3d31fd4cd7cc909fbeb310c5a75efdeb6291f6a98ee258851ba" Dec 03 13:51:03 crc kubenswrapper[4990]: I1203 13:51:03.286346 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:51:03 crc kubenswrapper[4990]: I1203 13:51:03.287038 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:51:03 crc kubenswrapper[4990]: I1203 13:51:03.287087 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:51:03 crc kubenswrapper[4990]: I1203 13:51:03.287875 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:51:03 crc kubenswrapper[4990]: I1203 13:51:03.287955 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" gracePeriod=600 Dec 03 13:51:04 crc kubenswrapper[4990]: E1203 13:51:04.105972 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:51:04 crc kubenswrapper[4990]: I1203 13:51:04.521082 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" exitCode=0 Dec 03 13:51:04 crc kubenswrapper[4990]: I1203 13:51:04.521414 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f"} Dec 03 13:51:04 crc kubenswrapper[4990]: I1203 13:51:04.521473 4990 scope.go:117] "RemoveContainer" containerID="33931fd5dcd79ed868236bde132bdc93fe9fd197d027f2241b26f817e20a5e50" Dec 03 13:51:04 crc kubenswrapper[4990]: I1203 13:51:04.522200 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:51:04 crc kubenswrapper[4990]: E1203 13:51:04.522553 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:51:18 crc kubenswrapper[4990]: I1203 13:51:18.264561 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:51:18 crc kubenswrapper[4990]: E1203 13:51:18.265220 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:51:32 crc kubenswrapper[4990]: I1203 13:51:32.273093 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:51:32 crc kubenswrapper[4990]: E1203 13:51:32.273784 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:51:44 crc kubenswrapper[4990]: I1203 13:51:44.265930 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:51:44 crc kubenswrapper[4990]: E1203 13:51:44.267526 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:51:58 crc kubenswrapper[4990]: I1203 13:51:58.265129 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:51:58 crc kubenswrapper[4990]: E1203 13:51:58.265878 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:52:11 crc kubenswrapper[4990]: I1203 13:52:11.268649 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:52:11 crc kubenswrapper[4990]: E1203 13:52:11.269792 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:52:26 crc kubenswrapper[4990]: I1203 13:52:26.266343 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:52:26 crc kubenswrapper[4990]: E1203 13:52:26.267169 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:52:27 crc kubenswrapper[4990]: I1203 13:52:27.390188 4990 scope.go:117] "RemoveContainer" containerID="904e528b2f302ad44050994e6c6aac295215a3d24aae9530ab47dee372a02927" Dec 03 13:52:37 crc kubenswrapper[4990]: I1203 13:52:37.265082 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:52:37 crc kubenswrapper[4990]: E1203 13:52:37.265894 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:52:52 crc kubenswrapper[4990]: I1203 13:52:52.275784 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:52:52 crc kubenswrapper[4990]: E1203 13:52:52.276799 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.160112 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4xm4x"] Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161095 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="extract-content" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161108 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="extract-content" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161120 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="registry-server" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161127 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="registry-server" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161142 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="extract-content" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161149 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="extract-content" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161165 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="extract-utilities" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161171 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="extract-utilities" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161184 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="gather" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161190 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="gather" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161204 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="extract-utilities" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161210 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="extract-utilities" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161220 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="registry-server" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161226 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="registry-server" Dec 03 13:52:58 crc kubenswrapper[4990]: E1203 13:52:58.161237 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="copy" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161243 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="copy" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161465 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc074cc5-41a4-41f4-bde6-d51f7eb5a1a4" containerName="registry-server" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161480 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="copy" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161496 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="21b0df2d-86cc-4e00-b4f5-5b9c2cbd62a0" containerName="registry-server" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.161508 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="741c2504-9580-4514-bf3a-0de631f99c03" containerName="gather" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.162966 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.178754 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xm4x"] Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.286000 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-catalog-content\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.286083 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwmp5\" (UniqueName: \"kubernetes.io/projected/266394d9-7e83-4a03-b6e9-d8a8b3695950-kube-api-access-qwmp5\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.286638 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-utilities\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.388709 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-utilities\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.388782 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-catalog-content\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.388825 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwmp5\" (UniqueName: \"kubernetes.io/projected/266394d9-7e83-4a03-b6e9-d8a8b3695950-kube-api-access-qwmp5\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.389330 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-utilities\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.389387 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-catalog-content\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.414558 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwmp5\" (UniqueName: \"kubernetes.io/projected/266394d9-7e83-4a03-b6e9-d8a8b3695950-kube-api-access-qwmp5\") pod \"redhat-marketplace-4xm4x\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.492970 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:52:58 crc kubenswrapper[4990]: I1203 13:52:58.962213 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xm4x"] Dec 03 13:52:59 crc kubenswrapper[4990]: I1203 13:52:59.551328 4990 generic.go:334] "Generic (PLEG): container finished" podID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerID="6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93" exitCode=0 Dec 03 13:52:59 crc kubenswrapper[4990]: I1203 13:52:59.551390 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xm4x" event={"ID":"266394d9-7e83-4a03-b6e9-d8a8b3695950","Type":"ContainerDied","Data":"6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93"} Dec 03 13:52:59 crc kubenswrapper[4990]: I1203 13:52:59.551669 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xm4x" event={"ID":"266394d9-7e83-4a03-b6e9-d8a8b3695950","Type":"ContainerStarted","Data":"bb3fe83cdfb8299532f4f88a452ad8446fa62dc6756edc9a1bcaf7d32ba53004"} Dec 03 13:52:59 crc kubenswrapper[4990]: I1203 13:52:59.553345 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:53:01 crc kubenswrapper[4990]: I1203 13:53:01.571789 4990 generic.go:334] "Generic (PLEG): container finished" podID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerID="a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2" exitCode=0 Dec 03 13:53:01 crc kubenswrapper[4990]: I1203 13:53:01.571905 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xm4x" event={"ID":"266394d9-7e83-4a03-b6e9-d8a8b3695950","Type":"ContainerDied","Data":"a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2"} Dec 03 13:53:03 crc kubenswrapper[4990]: I1203 13:53:03.593957 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xm4x" event={"ID":"266394d9-7e83-4a03-b6e9-d8a8b3695950","Type":"ContainerStarted","Data":"7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0"} Dec 03 13:53:03 crc kubenswrapper[4990]: I1203 13:53:03.622202 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4xm4x" podStartSLOduration=2.8542299399999997 podStartE2EDuration="5.622182929s" podCreationTimestamp="2025-12-03 13:52:58 +0000 UTC" firstStartedPulling="2025-12-03 13:52:59.553127798 +0000 UTC m=+4527.695039027" lastFinishedPulling="2025-12-03 13:53:02.321080787 +0000 UTC m=+4530.462992016" observedRunningTime="2025-12-03 13:53:03.611082823 +0000 UTC m=+4531.752994052" watchObservedRunningTime="2025-12-03 13:53:03.622182929 +0000 UTC m=+4531.764094158" Dec 03 13:53:07 crc kubenswrapper[4990]: I1203 13:53:07.265059 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:53:07 crc kubenswrapper[4990]: E1203 13:53:07.265862 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:53:08 crc kubenswrapper[4990]: I1203 13:53:08.494016 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:53:08 crc kubenswrapper[4990]: I1203 13:53:08.494560 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:53:08 crc kubenswrapper[4990]: I1203 13:53:08.544974 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:53:08 crc kubenswrapper[4990]: I1203 13:53:08.685869 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:53:09 crc kubenswrapper[4990]: I1203 13:53:09.554153 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xm4x"] Dec 03 13:53:10 crc kubenswrapper[4990]: I1203 13:53:10.656460 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4xm4x" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="registry-server" containerID="cri-o://7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0" gracePeriod=2 Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.185100 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.346671 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwmp5\" (UniqueName: \"kubernetes.io/projected/266394d9-7e83-4a03-b6e9-d8a8b3695950-kube-api-access-qwmp5\") pod \"266394d9-7e83-4a03-b6e9-d8a8b3695950\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.346822 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-utilities\") pod \"266394d9-7e83-4a03-b6e9-d8a8b3695950\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.346917 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-catalog-content\") pod \"266394d9-7e83-4a03-b6e9-d8a8b3695950\" (UID: \"266394d9-7e83-4a03-b6e9-d8a8b3695950\") " Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.347692 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-utilities" (OuterVolumeSpecName: "utilities") pod "266394d9-7e83-4a03-b6e9-d8a8b3695950" (UID: "266394d9-7e83-4a03-b6e9-d8a8b3695950"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.347805 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.351658 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/266394d9-7e83-4a03-b6e9-d8a8b3695950-kube-api-access-qwmp5" (OuterVolumeSpecName: "kube-api-access-qwmp5") pod "266394d9-7e83-4a03-b6e9-d8a8b3695950" (UID: "266394d9-7e83-4a03-b6e9-d8a8b3695950"). InnerVolumeSpecName "kube-api-access-qwmp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.368874 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "266394d9-7e83-4a03-b6e9-d8a8b3695950" (UID: "266394d9-7e83-4a03-b6e9-d8a8b3695950"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.449691 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266394d9-7e83-4a03-b6e9-d8a8b3695950-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.449755 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwmp5\" (UniqueName: \"kubernetes.io/projected/266394d9-7e83-4a03-b6e9-d8a8b3695950-kube-api-access-qwmp5\") on node \"crc\" DevicePath \"\"" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.670641 4990 generic.go:334] "Generic (PLEG): container finished" podID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerID="7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0" exitCode=0 Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.670688 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xm4x" event={"ID":"266394d9-7e83-4a03-b6e9-d8a8b3695950","Type":"ContainerDied","Data":"7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0"} Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.670704 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xm4x" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.670727 4990 scope.go:117] "RemoveContainer" containerID="7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.670716 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xm4x" event={"ID":"266394d9-7e83-4a03-b6e9-d8a8b3695950","Type":"ContainerDied","Data":"bb3fe83cdfb8299532f4f88a452ad8446fa62dc6756edc9a1bcaf7d32ba53004"} Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.692509 4990 scope.go:117] "RemoveContainer" containerID="a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.739050 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xm4x"] Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.747795 4990 scope.go:117] "RemoveContainer" containerID="6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.752228 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xm4x"] Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.795702 4990 scope.go:117] "RemoveContainer" containerID="7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0" Dec 03 13:53:11 crc kubenswrapper[4990]: E1203 13:53:11.796390 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0\": container with ID starting with 7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0 not found: ID does not exist" containerID="7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.796452 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0"} err="failed to get container status \"7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0\": rpc error: code = NotFound desc = could not find container \"7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0\": container with ID starting with 7116e22116acf38cd128e997faf63f9c989be3ad48715920869825caa5a6f4f0 not found: ID does not exist" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.796503 4990 scope.go:117] "RemoveContainer" containerID="a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2" Dec 03 13:53:11 crc kubenswrapper[4990]: E1203 13:53:11.796991 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2\": container with ID starting with a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2 not found: ID does not exist" containerID="a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.797054 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2"} err="failed to get container status \"a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2\": rpc error: code = NotFound desc = could not find container \"a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2\": container with ID starting with a92939aba4542eee4750af7ab9902cd94bbf770a968b325d46a2a68277c1fae2 not found: ID does not exist" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.797087 4990 scope.go:117] "RemoveContainer" containerID="6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93" Dec 03 13:53:11 crc kubenswrapper[4990]: E1203 13:53:11.797501 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93\": container with ID starting with 6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93 not found: ID does not exist" containerID="6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93" Dec 03 13:53:11 crc kubenswrapper[4990]: I1203 13:53:11.797528 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93"} err="failed to get container status \"6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93\": rpc error: code = NotFound desc = could not find container \"6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93\": container with ID starting with 6fc56700e33236ed85e69828d885c69786e601466400151c46e24e4e020add93 not found: ID does not exist" Dec 03 13:53:12 crc kubenswrapper[4990]: I1203 13:53:12.274888 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" path="/var/lib/kubelet/pods/266394d9-7e83-4a03-b6e9-d8a8b3695950/volumes" Dec 03 13:53:18 crc kubenswrapper[4990]: I1203 13:53:18.265067 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:53:18 crc kubenswrapper[4990]: E1203 13:53:18.265920 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:53:31 crc kubenswrapper[4990]: I1203 13:53:31.265499 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:53:31 crc kubenswrapper[4990]: E1203 13:53:31.266216 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:53:44 crc kubenswrapper[4990]: I1203 13:53:44.264863 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:53:44 crc kubenswrapper[4990]: E1203 13:53:44.265723 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.427983 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx545/must-gather-vkd6g"] Dec 03 13:53:47 crc kubenswrapper[4990]: E1203 13:53:47.428823 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="extract-utilities" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.428842 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="extract-utilities" Dec 03 13:53:47 crc kubenswrapper[4990]: E1203 13:53:47.428870 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="extract-content" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.428879 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="extract-content" Dec 03 13:53:47 crc kubenswrapper[4990]: E1203 13:53:47.428904 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="registry-server" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.428913 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="registry-server" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.429165 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="266394d9-7e83-4a03-b6e9-d8a8b3695950" containerName="registry-server" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.430415 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.433467 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zx545"/"openshift-service-ca.crt" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.433675 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zx545"/"default-dockercfg-pfbz9" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.433945 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zx545"/"kube-root-ca.crt" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.443586 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zx545/must-gather-vkd6g"] Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.575317 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/63813ce7-5ad3-45b3-a488-9de161b199c3-must-gather-output\") pod \"must-gather-vkd6g\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.575756 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-779q5\" (UniqueName: \"kubernetes.io/projected/63813ce7-5ad3-45b3-a488-9de161b199c3-kube-api-access-779q5\") pod \"must-gather-vkd6g\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.678044 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/63813ce7-5ad3-45b3-a488-9de161b199c3-must-gather-output\") pod \"must-gather-vkd6g\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.678193 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-779q5\" (UniqueName: \"kubernetes.io/projected/63813ce7-5ad3-45b3-a488-9de161b199c3-kube-api-access-779q5\") pod \"must-gather-vkd6g\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:47 crc kubenswrapper[4990]: I1203 13:53:47.678582 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/63813ce7-5ad3-45b3-a488-9de161b199c3-must-gather-output\") pod \"must-gather-vkd6g\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:48 crc kubenswrapper[4990]: I1203 13:53:48.081580 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-779q5\" (UniqueName: \"kubernetes.io/projected/63813ce7-5ad3-45b3-a488-9de161b199c3-kube-api-access-779q5\") pod \"must-gather-vkd6g\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:48 crc kubenswrapper[4990]: I1203 13:53:48.352959 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:53:48 crc kubenswrapper[4990]: I1203 13:53:48.847944 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zx545/must-gather-vkd6g"] Dec 03 13:53:48 crc kubenswrapper[4990]: W1203 13:53:48.852690 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63813ce7_5ad3_45b3_a488_9de161b199c3.slice/crio-ce77074a8ee3598d597559185d0469562b29ef7424e5e4787afaa3d5a1bd95d1 WatchSource:0}: Error finding container ce77074a8ee3598d597559185d0469562b29ef7424e5e4787afaa3d5a1bd95d1: Status 404 returned error can't find the container with id ce77074a8ee3598d597559185d0469562b29ef7424e5e4787afaa3d5a1bd95d1 Dec 03 13:53:49 crc kubenswrapper[4990]: I1203 13:53:49.072958 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/must-gather-vkd6g" event={"ID":"63813ce7-5ad3-45b3-a488-9de161b199c3","Type":"ContainerStarted","Data":"ce77074a8ee3598d597559185d0469562b29ef7424e5e4787afaa3d5a1bd95d1"} Dec 03 13:53:50 crc kubenswrapper[4990]: I1203 13:53:50.086471 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/must-gather-vkd6g" event={"ID":"63813ce7-5ad3-45b3-a488-9de161b199c3","Type":"ContainerStarted","Data":"47b87c38d1bce7629ddab6cac28acfb8c7278c0a60fdb840691374f1eba379b8"} Dec 03 13:53:51 crc kubenswrapper[4990]: I1203 13:53:51.095986 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/must-gather-vkd6g" event={"ID":"63813ce7-5ad3-45b3-a488-9de161b199c3","Type":"ContainerStarted","Data":"a0aa0f9019012eef297c823e52f6359c00f6f72e325e6bab90282fdd2b85b531"} Dec 03 13:53:51 crc kubenswrapper[4990]: I1203 13:53:51.114831 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zx545/must-gather-vkd6g" podStartSLOduration=4.114804798 podStartE2EDuration="4.114804798s" podCreationTimestamp="2025-12-03 13:53:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:51.110233016 +0000 UTC m=+4579.252144275" watchObservedRunningTime="2025-12-03 13:53:51.114804798 +0000 UTC m=+4579.256716037" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.491760 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx545/crc-debug-zhdws"] Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.493654 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.597733 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-host\") pod \"crc-debug-zhdws\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.598135 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ts8d\" (UniqueName: \"kubernetes.io/projected/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-kube-api-access-5ts8d\") pod \"crc-debug-zhdws\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.700529 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-host\") pod \"crc-debug-zhdws\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.700599 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ts8d\" (UniqueName: \"kubernetes.io/projected/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-kube-api-access-5ts8d\") pod \"crc-debug-zhdws\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.701114 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-host\") pod \"crc-debug-zhdws\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.723303 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ts8d\" (UniqueName: \"kubernetes.io/projected/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-kube-api-access-5ts8d\") pod \"crc-debug-zhdws\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: I1203 13:53:53.820246 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:53:53 crc kubenswrapper[4990]: W1203 13:53:53.877057 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b4972ba_a112_4f77_9cb0_be6ba7fa6a70.slice/crio-47295c8ed91b3bfd664dbc55c7ddfa6a4a4769449eacfdcbccd2c96027fc01cf WatchSource:0}: Error finding container 47295c8ed91b3bfd664dbc55c7ddfa6a4a4769449eacfdcbccd2c96027fc01cf: Status 404 returned error can't find the container with id 47295c8ed91b3bfd664dbc55c7ddfa6a4a4769449eacfdcbccd2c96027fc01cf Dec 03 13:53:54 crc kubenswrapper[4990]: I1203 13:53:54.126440 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-zhdws" event={"ID":"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70","Type":"ContainerStarted","Data":"47295c8ed91b3bfd664dbc55c7ddfa6a4a4769449eacfdcbccd2c96027fc01cf"} Dec 03 13:53:55 crc kubenswrapper[4990]: I1203 13:53:55.147070 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-zhdws" event={"ID":"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70","Type":"ContainerStarted","Data":"99dde99af72854895d0943426970d10b72e0fc1c528a197a7b0bb13f4328134c"} Dec 03 13:53:55 crc kubenswrapper[4990]: I1203 13:53:55.167524 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zx545/crc-debug-zhdws" podStartSLOduration=2.167508164 podStartE2EDuration="2.167508164s" podCreationTimestamp="2025-12-03 13:53:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 13:53:55.164929135 +0000 UTC m=+4583.306840364" watchObservedRunningTime="2025-12-03 13:53:55.167508164 +0000 UTC m=+4583.309419393" Dec 03 13:53:56 crc kubenswrapper[4990]: I1203 13:53:56.265336 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:53:56 crc kubenswrapper[4990]: E1203 13:53:56.266145 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:54:07 crc kubenswrapper[4990]: I1203 13:54:07.265183 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:54:07 crc kubenswrapper[4990]: E1203 13:54:07.266122 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:54:21 crc kubenswrapper[4990]: I1203 13:54:21.265020 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:54:21 crc kubenswrapper[4990]: E1203 13:54:21.265708 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:54:35 crc kubenswrapper[4990]: I1203 13:54:35.265404 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:54:35 crc kubenswrapper[4990]: E1203 13:54:35.266424 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:54:35 crc kubenswrapper[4990]: I1203 13:54:35.501382 4990 generic.go:334] "Generic (PLEG): container finished" podID="9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" containerID="99dde99af72854895d0943426970d10b72e0fc1c528a197a7b0bb13f4328134c" exitCode=0 Dec 03 13:54:35 crc kubenswrapper[4990]: I1203 13:54:35.501723 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-zhdws" event={"ID":"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70","Type":"ContainerDied","Data":"99dde99af72854895d0943426970d10b72e0fc1c528a197a7b0bb13f4328134c"} Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.627711 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.666445 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx545/crc-debug-zhdws"] Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.677150 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx545/crc-debug-zhdws"] Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.743828 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-host\") pod \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.743969 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-host" (OuterVolumeSpecName: "host") pod "9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" (UID: "9b4972ba-a112-4f77-9cb0-be6ba7fa6a70"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.744076 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ts8d\" (UniqueName: \"kubernetes.io/projected/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-kube-api-access-5ts8d\") pod \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\" (UID: \"9b4972ba-a112-4f77-9cb0-be6ba7fa6a70\") " Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.744520 4990 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-host\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.750087 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-kube-api-access-5ts8d" (OuterVolumeSpecName: "kube-api-access-5ts8d") pod "9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" (UID: "9b4972ba-a112-4f77-9cb0-be6ba7fa6a70"). InnerVolumeSpecName "kube-api-access-5ts8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:36 crc kubenswrapper[4990]: I1203 13:54:36.846545 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ts8d\" (UniqueName: \"kubernetes.io/projected/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70-kube-api-access-5ts8d\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.520636 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47295c8ed91b3bfd664dbc55c7ddfa6a4a4769449eacfdcbccd2c96027fc01cf" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.520704 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-zhdws" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.846800 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx545/crc-debug-8fcfg"] Dec 03 13:54:37 crc kubenswrapper[4990]: E1203 13:54:37.848195 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" containerName="container-00" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.848271 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" containerName="container-00" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.848559 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" containerName="container-00" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.849303 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.964442 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1c510f6-4162-4d2f-80c8-20dd062e68f3-host\") pod \"crc-debug-8fcfg\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:37 crc kubenswrapper[4990]: I1203 13:54:37.964564 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bwm8\" (UniqueName: \"kubernetes.io/projected/d1c510f6-4162-4d2f-80c8-20dd062e68f3-kube-api-access-4bwm8\") pod \"crc-debug-8fcfg\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:38 crc kubenswrapper[4990]: I1203 13:54:38.067023 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1c510f6-4162-4d2f-80c8-20dd062e68f3-host\") pod \"crc-debug-8fcfg\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:38 crc kubenswrapper[4990]: I1203 13:54:38.067108 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bwm8\" (UniqueName: \"kubernetes.io/projected/d1c510f6-4162-4d2f-80c8-20dd062e68f3-kube-api-access-4bwm8\") pod \"crc-debug-8fcfg\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:38 crc kubenswrapper[4990]: I1203 13:54:38.067208 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1c510f6-4162-4d2f-80c8-20dd062e68f3-host\") pod \"crc-debug-8fcfg\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:38 crc kubenswrapper[4990]: I1203 13:54:38.276514 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b4972ba-a112-4f77-9cb0-be6ba7fa6a70" path="/var/lib/kubelet/pods/9b4972ba-a112-4f77-9cb0-be6ba7fa6a70/volumes" Dec 03 13:54:38 crc kubenswrapper[4990]: I1203 13:54:38.383237 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bwm8\" (UniqueName: \"kubernetes.io/projected/d1c510f6-4162-4d2f-80c8-20dd062e68f3-kube-api-access-4bwm8\") pod \"crc-debug-8fcfg\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:38 crc kubenswrapper[4990]: I1203 13:54:38.464959 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:39 crc kubenswrapper[4990]: I1203 13:54:39.539615 4990 generic.go:334] "Generic (PLEG): container finished" podID="d1c510f6-4162-4d2f-80c8-20dd062e68f3" containerID="1b6cf0fec6d91bf7021ba4c617bfaa7a8c29b4f67f88c063e87b57080d5b9715" exitCode=0 Dec 03 13:54:39 crc kubenswrapper[4990]: I1203 13:54:39.540130 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-8fcfg" event={"ID":"d1c510f6-4162-4d2f-80c8-20dd062e68f3","Type":"ContainerDied","Data":"1b6cf0fec6d91bf7021ba4c617bfaa7a8c29b4f67f88c063e87b57080d5b9715"} Dec 03 13:54:39 crc kubenswrapper[4990]: I1203 13:54:39.540190 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-8fcfg" event={"ID":"d1c510f6-4162-4d2f-80c8-20dd062e68f3","Type":"ContainerStarted","Data":"65e57b69f7df9167c7d82bd1b6f1883d8f64223add6fedcf497cc7f48d3fcb64"} Dec 03 13:54:39 crc kubenswrapper[4990]: I1203 13:54:39.978498 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx545/crc-debug-8fcfg"] Dec 03 13:54:39 crc kubenswrapper[4990]: I1203 13:54:39.989886 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx545/crc-debug-8fcfg"] Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.656168 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.819156 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bwm8\" (UniqueName: \"kubernetes.io/projected/d1c510f6-4162-4d2f-80c8-20dd062e68f3-kube-api-access-4bwm8\") pod \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.819397 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1c510f6-4162-4d2f-80c8-20dd062e68f3-host\") pod \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\" (UID: \"d1c510f6-4162-4d2f-80c8-20dd062e68f3\") " Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.819569 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1c510f6-4162-4d2f-80c8-20dd062e68f3-host" (OuterVolumeSpecName: "host") pod "d1c510f6-4162-4d2f-80c8-20dd062e68f3" (UID: "d1c510f6-4162-4d2f-80c8-20dd062e68f3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.819874 4990 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1c510f6-4162-4d2f-80c8-20dd062e68f3-host\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.827701 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1c510f6-4162-4d2f-80c8-20dd062e68f3-kube-api-access-4bwm8" (OuterVolumeSpecName: "kube-api-access-4bwm8") pod "d1c510f6-4162-4d2f-80c8-20dd062e68f3" (UID: "d1c510f6-4162-4d2f-80c8-20dd062e68f3"). InnerVolumeSpecName "kube-api-access-4bwm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:40 crc kubenswrapper[4990]: I1203 13:54:40.921720 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bwm8\" (UniqueName: \"kubernetes.io/projected/d1c510f6-4162-4d2f-80c8-20dd062e68f3-kube-api-access-4bwm8\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.440165 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx545/crc-debug-4ltjz"] Dec 03 13:54:41 crc kubenswrapper[4990]: E1203 13:54:41.440599 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c510f6-4162-4d2f-80c8-20dd062e68f3" containerName="container-00" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.440618 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c510f6-4162-4d2f-80c8-20dd062e68f3" containerName="container-00" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.440785 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1c510f6-4162-4d2f-80c8-20dd062e68f3" containerName="container-00" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.441440 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.531653 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzw2g\" (UniqueName: \"kubernetes.io/projected/28bafa19-910a-4a32-97e5-89def6c7628b-kube-api-access-vzw2g\") pod \"crc-debug-4ltjz\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.531778 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28bafa19-910a-4a32-97e5-89def6c7628b-host\") pod \"crc-debug-4ltjz\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.560861 4990 scope.go:117] "RemoveContainer" containerID="1b6cf0fec6d91bf7021ba4c617bfaa7a8c29b4f67f88c063e87b57080d5b9715" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.560958 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-8fcfg" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.633753 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28bafa19-910a-4a32-97e5-89def6c7628b-host\") pod \"crc-debug-4ltjz\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.633909 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28bafa19-910a-4a32-97e5-89def6c7628b-host\") pod \"crc-debug-4ltjz\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.634015 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzw2g\" (UniqueName: \"kubernetes.io/projected/28bafa19-910a-4a32-97e5-89def6c7628b-kube-api-access-vzw2g\") pod \"crc-debug-4ltjz\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.662956 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzw2g\" (UniqueName: \"kubernetes.io/projected/28bafa19-910a-4a32-97e5-89def6c7628b-kube-api-access-vzw2g\") pod \"crc-debug-4ltjz\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:41 crc kubenswrapper[4990]: I1203 13:54:41.759239 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:42 crc kubenswrapper[4990]: I1203 13:54:42.297516 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1c510f6-4162-4d2f-80c8-20dd062e68f3" path="/var/lib/kubelet/pods/d1c510f6-4162-4d2f-80c8-20dd062e68f3/volumes" Dec 03 13:54:42 crc kubenswrapper[4990]: I1203 13:54:42.571660 4990 generic.go:334] "Generic (PLEG): container finished" podID="28bafa19-910a-4a32-97e5-89def6c7628b" containerID="ccf9645424469955c1bcaeeaa607015eda02d4c08e74187d93f66f1e3711d846" exitCode=0 Dec 03 13:54:42 crc kubenswrapper[4990]: I1203 13:54:42.571763 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-4ltjz" event={"ID":"28bafa19-910a-4a32-97e5-89def6c7628b","Type":"ContainerDied","Data":"ccf9645424469955c1bcaeeaa607015eda02d4c08e74187d93f66f1e3711d846"} Dec 03 13:54:42 crc kubenswrapper[4990]: I1203 13:54:42.571799 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/crc-debug-4ltjz" event={"ID":"28bafa19-910a-4a32-97e5-89def6c7628b","Type":"ContainerStarted","Data":"ccc689d3a4f4c9d0c6db6130ce7cb145d965fc1686ef3e1bd81d2dd08f50fb52"} Dec 03 13:54:42 crc kubenswrapper[4990]: I1203 13:54:42.605841 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx545/crc-debug-4ltjz"] Dec 03 13:54:42 crc kubenswrapper[4990]: I1203 13:54:42.614826 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx545/crc-debug-4ltjz"] Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.701107 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.875287 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzw2g\" (UniqueName: \"kubernetes.io/projected/28bafa19-910a-4a32-97e5-89def6c7628b-kube-api-access-vzw2g\") pod \"28bafa19-910a-4a32-97e5-89def6c7628b\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.875532 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28bafa19-910a-4a32-97e5-89def6c7628b-host\") pod \"28bafa19-910a-4a32-97e5-89def6c7628b\" (UID: \"28bafa19-910a-4a32-97e5-89def6c7628b\") " Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.875653 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28bafa19-910a-4a32-97e5-89def6c7628b-host" (OuterVolumeSpecName: "host") pod "28bafa19-910a-4a32-97e5-89def6c7628b" (UID: "28bafa19-910a-4a32-97e5-89def6c7628b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.876140 4990 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28bafa19-910a-4a32-97e5-89def6c7628b-host\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.882207 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28bafa19-910a-4a32-97e5-89def6c7628b-kube-api-access-vzw2g" (OuterVolumeSpecName: "kube-api-access-vzw2g") pod "28bafa19-910a-4a32-97e5-89def6c7628b" (UID: "28bafa19-910a-4a32-97e5-89def6c7628b"). InnerVolumeSpecName "kube-api-access-vzw2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:54:43 crc kubenswrapper[4990]: I1203 13:54:43.978490 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzw2g\" (UniqueName: \"kubernetes.io/projected/28bafa19-910a-4a32-97e5-89def6c7628b-kube-api-access-vzw2g\") on node \"crc\" DevicePath \"\"" Dec 03 13:54:44 crc kubenswrapper[4990]: I1203 13:54:44.276285 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28bafa19-910a-4a32-97e5-89def6c7628b" path="/var/lib/kubelet/pods/28bafa19-910a-4a32-97e5-89def6c7628b/volumes" Dec 03 13:54:44 crc kubenswrapper[4990]: I1203 13:54:44.589869 4990 scope.go:117] "RemoveContainer" containerID="ccf9645424469955c1bcaeeaa607015eda02d4c08e74187d93f66f1e3711d846" Dec 03 13:54:44 crc kubenswrapper[4990]: I1203 13:54:44.589921 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/crc-debug-4ltjz" Dec 03 13:54:49 crc kubenswrapper[4990]: I1203 13:54:49.264358 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:54:49 crc kubenswrapper[4990]: E1203 13:54:49.265052 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:55:00 crc kubenswrapper[4990]: I1203 13:55:00.533465 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65cfc74cf4-n77zl_630b8058-02f8-4c7e-aeb7-6cca035356ed/barbican-api/0.log" Dec 03 13:55:00 crc kubenswrapper[4990]: I1203 13:55:00.704929 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65cfc74cf4-n77zl_630b8058-02f8-4c7e-aeb7-6cca035356ed/barbican-api-log/0.log" Dec 03 13:55:00 crc kubenswrapper[4990]: I1203 13:55:00.784691 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68db7d64bb-tmc56_fc68490b-9eed-4344-a09f-96adddd6c7c7/barbican-keystone-listener/0.log" Dec 03 13:55:00 crc kubenswrapper[4990]: I1203 13:55:00.824044 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68db7d64bb-tmc56_fc68490b-9eed-4344-a09f-96adddd6c7c7/barbican-keystone-listener-log/0.log" Dec 03 13:55:00 crc kubenswrapper[4990]: I1203 13:55:00.965701 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-67f9cc6bf5-jfjbk_8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b/barbican-worker/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.046181 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-67f9cc6bf5-jfjbk_8914ad61-3a71-4ec2-bf91-b3d0f69cfa1b/barbican-worker-log/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.181257 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rx2d7_6c0707b9-5be8-4bf4-870b-ace30c9a9fe0/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.249209 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/ceilometer-central-agent/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.356827 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/ceilometer-notification-agent/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.381268 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/sg-core/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.411981 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_901bdc2a-da65-4388-9865-67a1ff3eec17/proxy-httpd/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.811193 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8534a0f8-e17f-49e3-948f-475e422d43e8/cinder-api/0.log" Dec 03 13:55:01 crc kubenswrapper[4990]: I1203 13:55:01.823779 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8534a0f8-e17f-49e3-948f-475e422d43e8/cinder-api-log/0.log" Dec 03 13:55:02 crc kubenswrapper[4990]: I1203 13:55:02.516895 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d724111e-b7f1-4efe-a05e-1e5cf2f5a278/probe/0.log" Dec 03 13:55:02 crc kubenswrapper[4990]: I1203 13:55:02.556833 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wltmb_7c2c7e25-8434-436c-8e35-61011e98873c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:02 crc kubenswrapper[4990]: I1203 13:55:02.591415 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d724111e-b7f1-4efe-a05e-1e5cf2f5a278/cinder-scheduler/0.log" Dec 03 13:55:02 crc kubenswrapper[4990]: I1203 13:55:02.839582 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tf6pm_8c311218-88d9-4022-8962-8ff68d03fa29/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:02 crc kubenswrapper[4990]: I1203 13:55:02.895697 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-s9vgm_f65f22f9-af79-4a63-bfc8-590d88507cb3/init/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.056027 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-s9vgm_f65f22f9-af79-4a63-bfc8-590d88507cb3/init/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.090348 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-s9vgm_f65f22f9-af79-4a63-bfc8-590d88507cb3/dnsmasq-dns/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.104188 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-57tqr_48109f08-9566-4bb0-bc9c-68413ddd32ae/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.265582 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:55:03 crc kubenswrapper[4990]: E1203 13:55:03.265837 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.322899 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fbdb7170-e955-4fd4-bfd5-03ad13363aa9/glance-httpd/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.341198 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fbdb7170-e955-4fd4-bfd5-03ad13363aa9/glance-log/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.516920 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_fd733a23-295d-4c67-9f67-1e27d99d7847/glance-httpd/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.539693 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_fd733a23-295d-4c67-9f67-1e27d99d7847/glance-log/0.log" Dec 03 13:55:03 crc kubenswrapper[4990]: I1203 13:55:03.714001 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5f6447d884-4db4z_c626101c-4745-40ce-aba1-587cb02ea499/horizon/0.log" Dec 03 13:55:04 crc kubenswrapper[4990]: I1203 13:55:04.486460 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6cm2v_213d788e-4e19-4020-a8ef-1d84d287fb95/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:04 crc kubenswrapper[4990]: I1203 13:55:04.515660 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-blhdl_a777faab-eccb-4a4f-b181-836edc76d745/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:04 crc kubenswrapper[4990]: I1203 13:55:04.873429 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5f6447d884-4db4z_c626101c-4745-40ce-aba1-587cb02ea499/horizon-log/0.log" Dec 03 13:55:04 crc kubenswrapper[4990]: I1203 13:55:04.897542 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b60c63f3-9d80-4660-9bc9-4afef4bd7cd6/kube-state-metrics/0.log" Dec 03 13:55:05 crc kubenswrapper[4990]: I1203 13:55:05.004324 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6d44745d6d-9zbl5_6e4c5463-6d0f-4e0f-bfc3-e8e04cfab383/keystone-api/0.log" Dec 03 13:55:05 crc kubenswrapper[4990]: I1203 13:55:05.120996 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9qzpk_4c4068fd-3e0a-442a-b035-f2e84fba7e88/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:05 crc kubenswrapper[4990]: I1203 13:55:05.461871 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b4f6458f-phdfw_a486143f-bd72-4292-aaa1-194fb374862a/neutron-api/0.log" Dec 03 13:55:05 crc kubenswrapper[4990]: I1203 13:55:05.473041 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b4f6458f-phdfw_a486143f-bd72-4292-aaa1-194fb374862a/neutron-httpd/0.log" Dec 03 13:55:05 crc kubenswrapper[4990]: I1203 13:55:05.549416 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gk42w_52371a09-a7eb-42bb-aa53-3a2188c6658d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:06 crc kubenswrapper[4990]: I1203 13:55:06.216677 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c704a76d-fc53-4741-9bd5-b7893e1c96a5/nova-api-log/0.log" Dec 03 13:55:06 crc kubenswrapper[4990]: I1203 13:55:06.276641 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_11fb2cd9-0846-413e-b00a-29f258f641c5/nova-cell0-conductor-conductor/0.log" Dec 03 13:55:06 crc kubenswrapper[4990]: I1203 13:55:06.530956 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_a34660cc-b629-4ac1-ab68-0546243d600a/nova-cell1-conductor-conductor/0.log" Dec 03 13:55:06 crc kubenswrapper[4990]: I1203 13:55:06.708764 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_37a5c5c8-e837-456a-94ac-bd2d436b7e44/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 13:55:06 crc kubenswrapper[4990]: I1203 13:55:06.725185 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c704a76d-fc53-4741-9bd5-b7893e1c96a5/nova-api-api/0.log" Dec 03 13:55:06 crc kubenswrapper[4990]: I1203 13:55:06.801276 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hhvts_4191ee3d-7722-4690-938c-7b8d8478589e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:07 crc kubenswrapper[4990]: I1203 13:55:07.065666 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f4737b59-9b3b-4bd5-8220-1346a05c998f/nova-metadata-log/0.log" Dec 03 13:55:07 crc kubenswrapper[4990]: I1203 13:55:07.391996 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81e14867-9f93-4820-ba0f-c6dbd12e31db/mysql-bootstrap/0.log" Dec 03 13:55:07 crc kubenswrapper[4990]: I1203 13:55:07.489200 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_9e634575-20df-49bc-b2af-5322b408b702/nova-scheduler-scheduler/0.log" Dec 03 13:55:07 crc kubenswrapper[4990]: I1203 13:55:07.617566 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81e14867-9f93-4820-ba0f-c6dbd12e31db/mysql-bootstrap/0.log" Dec 03 13:55:07 crc kubenswrapper[4990]: I1203 13:55:07.620613 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81e14867-9f93-4820-ba0f-c6dbd12e31db/galera/0.log" Dec 03 13:55:07 crc kubenswrapper[4990]: I1203 13:55:07.818996 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4fb6a200-beee-4b47-ac8d-e370e307070f/mysql-bootstrap/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.076812 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4fb6a200-beee-4b47-ac8d-e370e307070f/mysql-bootstrap/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.131716 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_4fb6a200-beee-4b47-ac8d-e370e307070f/galera/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.273962 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9fdd0021-daac-4e15-b565-25cc720ed808/openstackclient/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.430401 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6frjn_83ad84ef-dc0d-4b8e-9efc-21d39e7c99e5/ovn-controller/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.549177 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-dpmrl_caff07ff-f45e-4438-96a7-545d4de585c0/openstack-network-exporter/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.645842 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f4737b59-9b3b-4bd5-8220-1346a05c998f/nova-metadata-metadata/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.744537 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovsdb-server-init/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.962286 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovsdb-server-init/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.973212 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovs-vswitchd/0.log" Dec 03 13:55:08 crc kubenswrapper[4990]: I1203 13:55:08.999760 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lbt2k_3f81fd15-3954-40f5-9033-18cb8e9b01ed/ovsdb-server/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.254343 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-q8c6t_3cf5d887-0878-407b-98c9-539cb967a73e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.289296 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7d581ac6-ba40-4773-9d18-5aa2fcdfad62/openstack-network-exporter/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.311626 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7d581ac6-ba40-4773-9d18-5aa2fcdfad62/ovn-northd/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.504610 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8ae85bd2-714b-4a62-b287-c2a9780753bd/openstack-network-exporter/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.588807 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8ae85bd2-714b-4a62-b287-c2a9780753bd/ovsdbserver-nb/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.691389 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_63ff1335-eac5-4524-b448-785488679bff/openstack-network-exporter/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.768993 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_63ff1335-eac5-4524-b448-785488679bff/ovsdbserver-sb/0.log" Dec 03 13:55:09 crc kubenswrapper[4990]: I1203 13:55:09.945979 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-644b9ffb4d-v4tbk_11b365c1-27dc-4c2f-a79c-12069d221f7f/placement-api/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.094092 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-644b9ffb4d-v4tbk_11b365c1-27dc-4c2f-a79c-12069d221f7f/placement-log/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.114835 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_be4855bd-7c56-4d1c-af0d-4017083409a6/setup-container/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.323003 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98e2d7f8-4691-44ca-a46f-9e08c82844a2/setup-container/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.408056 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_be4855bd-7c56-4d1c-af0d-4017083409a6/rabbitmq/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.409229 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_be4855bd-7c56-4d1c-af0d-4017083409a6/setup-container/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.692419 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98e2d7f8-4691-44ca-a46f-9e08c82844a2/setup-container/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.695590 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98e2d7f8-4691-44ca-a46f-9e08c82844a2/rabbitmq/0.log" Dec 03 13:55:10 crc kubenswrapper[4990]: I1203 13:55:10.736474 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-btf6p_5398edd4-615e-4e57-b25b-e7732ef5dbf3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.007693 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-d2cf7_748e28f7-7642-4591-bdba-29a79d86f5af/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.046318 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p972z_3c11a50b-b2d3-45be-9999-bcd93b698b1c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.314784 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-hbnrv_58dc37ce-d8bf-46f0-9338-43620480d0dd/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.326680 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ntk2h_dfc3e1a9-2b0f-4242-9e84-ca8332cca9cf/ssh-known-hosts-edpm-deployment/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.581959 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86d5555b99-xbxkx_62fcb6e6-b158-4a76-8570-3563346e2a77/proxy-server/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.767210 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86d5555b99-xbxkx_62fcb6e6-b158-4a76-8570-3563346e2a77/proxy-httpd/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.801946 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gqdz2_73f3c670-ab5f-4ab9-9278-9aa008c9a35b/swift-ring-rebalance/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.987075 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-auditor/0.log" Dec 03 13:55:11 crc kubenswrapper[4990]: I1203 13:55:11.993165 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-reaper/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.149600 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-replicator/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.234385 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-auditor/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.267749 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/account-server/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.287805 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-replicator/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.774194 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-server/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.868794 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/container-updater/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.881990 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-expirer/0.log" Dec 03 13:55:12 crc kubenswrapper[4990]: I1203 13:55:12.883581 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-auditor/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.043668 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-replicator/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.104834 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/rsync/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.105664 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-server/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.137302 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/object-updater/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.276870 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f7d434c5-0459-4d5a-a401-4a0c4b82a553/swift-recon-cron/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.483370 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zkgn9_882b0d7e-4fd5-4462-ba56-091dbc993c0b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.550989 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_f375426e-6095-4621-ab29-6c7324da86e7/tempest-tests-tempest-tests-runner/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.608896 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f0fab240-b45f-44e1-8fb7-446c4dc6d64e/test-operator-logs-container/0.log" Dec 03 13:55:13 crc kubenswrapper[4990]: I1203 13:55:13.806030 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wqzxv_5207255c-c5b2-470d-aa44-57da1579fe01/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 13:55:14 crc kubenswrapper[4990]: I1203 13:55:14.264316 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:55:14 crc kubenswrapper[4990]: E1203 13:55:14.264649 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:55:25 crc kubenswrapper[4990]: I1203 13:55:25.884104 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_24350652-92e4-40b6-82d5-4b785a2a0314/memcached/0.log" Dec 03 13:55:29 crc kubenswrapper[4990]: I1203 13:55:29.265009 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:55:29 crc kubenswrapper[4990]: E1203 13:55:29.265898 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:55:40 crc kubenswrapper[4990]: I1203 13:55:40.723531 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4sp89_c7a7a3b2-d356-4dfa-b640-ae39aa2d4840/kube-rbac-proxy/0.log" Dec 03 13:55:40 crc kubenswrapper[4990]: I1203 13:55:40.855001 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4sp89_c7a7a3b2-d356-4dfa-b640-ae39aa2d4840/manager/0.log" Dec 03 13:55:40 crc kubenswrapper[4990]: I1203 13:55:40.918372 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7n65k_c45d3b36-d188-4bb1-b19f-5521709f572a/kube-rbac-proxy/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.083980 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7n65k_c45d3b36-d188-4bb1-b19f-5521709f572a/manager/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.140295 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hqhfc_93f0e562-5973-4c2a-ab48-aa167ae49ffa/kube-rbac-proxy/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.210529 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-hqhfc_93f0e562-5973-4c2a-ab48-aa167ae49ffa/manager/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.275729 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/util/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.436856 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/pull/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.519580 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/pull/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.527780 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/util/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.643545 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/util/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.666198 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/pull/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.709999 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decgd8hm_50576f83-b05b-418b-a57d-0a322f2c489f/extract/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.818600 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-z484w_ac715eff-4f7f-44b0-b5ca-c5d283edca36/kube-rbac-proxy/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.910122 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-z484w_ac715eff-4f7f-44b0-b5ca-c5d283edca36/manager/0.log" Dec 03 13:55:41 crc kubenswrapper[4990]: I1203 13:55:41.955183 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xt2lq_a8fb4bb3-4b67-4ada-8876-389bff0e9836/kube-rbac-proxy/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.027463 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xt2lq_a8fb4bb3-4b67-4ada-8876-389bff0e9836/manager/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.152559 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-8zwv5_4881a1ed-a968-4f20-a8e6-94a91ec5eceb/kube-rbac-proxy/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.184154 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-8zwv5_4881a1ed-a968-4f20-a8e6-94a91ec5eceb/manager/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.344220 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zhgnf_0e880282-b396-4fa9-a3ac-89d81315ecd8/kube-rbac-proxy/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.538537 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-79xxb_96da719f-5659-45eb-ac48-37165c936527/kube-rbac-proxy/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.571933 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zhgnf_0e880282-b396-4fa9-a3ac-89d81315ecd8/manager/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.596008 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-79xxb_96da719f-5659-45eb-ac48-37165c936527/manager/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.727111 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-256rj_43c2b7af-9659-4ae6-8809-27475b31c611/kube-rbac-proxy/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.857966 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-256rj_43c2b7af-9659-4ae6-8809-27475b31c611/manager/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.950349 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qt45w_9106e41f-5788-46a2-86ce-6362832d6df7/kube-rbac-proxy/0.log" Dec 03 13:55:42 crc kubenswrapper[4990]: I1203 13:55:42.968253 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qt45w_9106e41f-5788-46a2-86ce-6362832d6df7/manager/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.065305 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vkrn9_e944cb60-4e1a-448f-ab25-60ff87e3a166/kube-rbac-proxy/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.173806 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vkrn9_e944cb60-4e1a-448f-ab25-60ff87e3a166/manager/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.265380 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:55:43 crc kubenswrapper[4990]: E1203 13:55:43.265774 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.266269 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-c8cgl_0a2be3dd-196c-4186-ac5d-85ce550aa201/kube-rbac-proxy/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.326379 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-c8cgl_0a2be3dd-196c-4186-ac5d-85ce550aa201/manager/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.421536 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dg8m2_a18809dc-c194-44b4-84e8-26e6da515bb7/kube-rbac-proxy/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.541026 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dg8m2_a18809dc-c194-44b4-84e8-26e6da515bb7/manager/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.628498 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-xg94h_6ca15749-37c2-477f-8a9a-33c07b97aaf3/kube-rbac-proxy/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.666104 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-xg94h_6ca15749-37c2-477f-8a9a-33c07b97aaf3/manager/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.750168 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9_c0833459-1161-4beb-ad68-07d51b5b33d7/kube-rbac-proxy/0.log" Dec 03 13:55:43 crc kubenswrapper[4990]: I1203 13:55:43.824019 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd49fjg9_c0833459-1161-4beb-ad68-07d51b5b33d7/manager/0.log" Dec 03 13:55:44 crc kubenswrapper[4990]: I1203 13:55:44.289515 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xgxsw_c1b91baf-8945-490b-987c-531d89259d3b/registry-server/0.log" Dec 03 13:55:44 crc kubenswrapper[4990]: I1203 13:55:44.297095 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85bd6bf7bc-zdj92_49fd43e1-fcca-4a4e-a1fd-a6b320a11b0d/operator/0.log" Dec 03 13:55:44 crc kubenswrapper[4990]: I1203 13:55:44.575764 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-4db2x_64f072f3-940c-431d-8e0f-b77e9349e79e/kube-rbac-proxy/0.log" Dec 03 13:55:44 crc kubenswrapper[4990]: I1203 13:55:44.593941 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-4db2x_64f072f3-940c-431d-8e0f-b77e9349e79e/manager/0.log" Dec 03 13:55:44 crc kubenswrapper[4990]: I1203 13:55:44.805384 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-x7jtx_3443975f-d482-422a-a010-9940c2b1182f/kube-rbac-proxy/0.log" Dec 03 13:55:44 crc kubenswrapper[4990]: I1203 13:55:44.948362 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-x7jtx_3443975f-d482-422a-a010-9940c2b1182f/manager/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.241125 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5b474d9948-p5wjj_425132be-e04e-46f6-ac06-1546c5b12a26/manager/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.495936 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-p8hpt_cc2d7750-b3b5-4585-95fd-4f84e1c437ad/operator/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.512371 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-rvslq_8bf181cd-d802-43c1-a7f0-7df5ec5008a1/kube-rbac-proxy/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.514276 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-rvslq_8bf181cd-d802-43c1-a7f0-7df5ec5008a1/manager/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.677823 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-4wkxw_213fe9a3-0270-4ec5-b380-2f28159b5b6a/kube-rbac-proxy/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.783367 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-4wkxw_213fe9a3-0270-4ec5-b380-2f28159b5b6a/manager/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.872037 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tg4ww_4481645b-1e70-415a-a01b-131b1761dd19/kube-rbac-proxy/0.log" Dec 03 13:55:45 crc kubenswrapper[4990]: I1203 13:55:45.960055 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tg4ww_4481645b-1e70-415a-a01b-131b1761dd19/manager/0.log" Dec 03 13:55:46 crc kubenswrapper[4990]: I1203 13:55:46.034076 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7cl8r_d2c9d382-f60c-4c2c-a684-ddec4371b165/kube-rbac-proxy/0.log" Dec 03 13:55:46 crc kubenswrapper[4990]: I1203 13:55:46.063896 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7cl8r_d2c9d382-f60c-4c2c-a684-ddec4371b165/manager/0.log" Dec 03 13:55:54 crc kubenswrapper[4990]: I1203 13:55:54.264567 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:55:54 crc kubenswrapper[4990]: E1203 13:55:54.265414 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 13:56:06 crc kubenswrapper[4990]: I1203 13:56:06.264730 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:56:06 crc kubenswrapper[4990]: I1203 13:56:06.942712 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jsjk4_a36bdb31-4352-4ef2-9d56-85cd07b89e21/control-plane-machine-set-operator/0.log" Dec 03 13:56:07 crc kubenswrapper[4990]: I1203 13:56:07.159379 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jfgrs_184ef2da-fe68-4fbc-8a55-f6eb63262764/kube-rbac-proxy/0.log" Dec 03 13:56:07 crc kubenswrapper[4990]: I1203 13:56:07.377486 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"28c51c6562bf6ec0845cefe19910d3a93dce4c1ca0c2fd94d19a2e02f1f60412"} Dec 03 13:56:07 crc kubenswrapper[4990]: I1203 13:56:07.841860 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jfgrs_184ef2da-fe68-4fbc-8a55-f6eb63262764/machine-api-operator/0.log" Dec 03 13:56:21 crc kubenswrapper[4990]: I1203 13:56:21.095499 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xrqmv_555f5cb7-9649-4c3a-8b77-244edcb83c7f/cert-manager-controller/0.log" Dec 03 13:56:21 crc kubenswrapper[4990]: I1203 13:56:21.212780 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-xlfdw_9dd6b4f0-81be-480c-84d6-22ba5b462ab0/cert-manager-cainjector/0.log" Dec 03 13:56:21 crc kubenswrapper[4990]: I1203 13:56:21.293228 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8ghd6_3196eb4c-4adf-493f-ade6-a9c839d0e4e6/cert-manager-webhook/0.log" Dec 03 13:56:35 crc kubenswrapper[4990]: I1203 13:56:35.486470 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7n2ph_c717132b-5c00-4d42-9d18-c163e82b27ff/nmstate-console-plugin/0.log" Dec 03 13:56:35 crc kubenswrapper[4990]: I1203 13:56:35.512647 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dtt26_20a68e4a-3638-430c-a27b-7d535613c3d3/nmstate-handler/0.log" Dec 03 13:56:35 crc kubenswrapper[4990]: I1203 13:56:35.692267 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-gwfbg_afaa9abc-cdbc-44b9-ab3e-a69df533460c/kube-rbac-proxy/0.log" Dec 03 13:56:35 crc kubenswrapper[4990]: I1203 13:56:35.741438 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-gwfbg_afaa9abc-cdbc-44b9-ab3e-a69df533460c/nmstate-metrics/0.log" Dec 03 13:56:35 crc kubenswrapper[4990]: I1203 13:56:35.909037 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-g69mt_9429610d-e1cb-4248-8224-12e08acf1a69/nmstate-operator/0.log" Dec 03 13:56:35 crc kubenswrapper[4990]: I1203 13:56:35.945657 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-pgjpk_45d5180b-e7cb-481e-88c7-9fa11bf8edc9/nmstate-webhook/0.log" Dec 03 13:56:51 crc kubenswrapper[4990]: I1203 13:56:51.824246 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hhxt8_9d94113a-0245-4057-97e6-8464ef2a830d/kube-rbac-proxy/0.log" Dec 03 13:56:51 crc kubenswrapper[4990]: I1203 13:56:51.930643 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hhxt8_9d94113a-0245-4057-97e6-8464ef2a830d/controller/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.059923 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.256233 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.276989 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.278073 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.288749 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.448240 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.499025 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.504095 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.524708 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.662422 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-frr-files/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.678102 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-reloader/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.701604 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/cp-metrics/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.706174 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/controller/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.891542 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/kube-rbac-proxy-frr/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.913728 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/frr-metrics/0.log" Dec 03 13:56:52 crc kubenswrapper[4990]: I1203 13:56:52.914539 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/kube-rbac-proxy/0.log" Dec 03 13:56:53 crc kubenswrapper[4990]: I1203 13:56:53.107555 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-xnhk6_d64e42b3-c578-44d9-818b-3b0aa7f4b3e4/frr-k8s-webhook-server/0.log" Dec 03 13:56:53 crc kubenswrapper[4990]: I1203 13:56:53.142302 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/reloader/0.log" Dec 03 13:56:53 crc kubenswrapper[4990]: I1203 13:56:53.405345 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-584cf47849-n84tl_5aac7d67-caa8-481d-80fb-7ca848f1007f/manager/0.log" Dec 03 13:56:53 crc kubenswrapper[4990]: I1203 13:56:53.586344 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-8464f887b-s5zch_1d95a171-2b09-4441-b7e7-31a5f4427a50/webhook-server/0.log" Dec 03 13:56:53 crc kubenswrapper[4990]: I1203 13:56:53.601099 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-t6cl8_10554cf0-1dc9-4e97-973f-1f2d47e3d1f8/kube-rbac-proxy/0.log" Dec 03 13:56:54 crc kubenswrapper[4990]: I1203 13:56:54.254934 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-t6cl8_10554cf0-1dc9-4e97-973f-1f2d47e3d1f8/speaker/0.log" Dec 03 13:56:54 crc kubenswrapper[4990]: I1203 13:56:54.461665 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p4dt2_f6a369fa-06fe-4bb3-a796-c54b8442582e/frr/0.log" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.190620 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svrj2"] Dec 03 13:57:04 crc kubenswrapper[4990]: E1203 13:57:04.191774 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bafa19-910a-4a32-97e5-89def6c7628b" containerName="container-00" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.191794 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bafa19-910a-4a32-97e5-89def6c7628b" containerName="container-00" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.192028 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="28bafa19-910a-4a32-97e5-89def6c7628b" containerName="container-00" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.193876 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.224296 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svrj2"] Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.321749 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-utilities\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.324481 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-catalog-content\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.324741 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmk9d\" (UniqueName: \"kubernetes.io/projected/554c4066-d7ce-4bd5-be73-ebb33578a66c-kube-api-access-rmk9d\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.426412 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmk9d\" (UniqueName: \"kubernetes.io/projected/554c4066-d7ce-4bd5-be73-ebb33578a66c-kube-api-access-rmk9d\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.426551 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-utilities\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.426637 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-catalog-content\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.427127 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-catalog-content\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.427693 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-utilities\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.449623 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmk9d\" (UniqueName: \"kubernetes.io/projected/554c4066-d7ce-4bd5-be73-ebb33578a66c-kube-api-access-rmk9d\") pod \"community-operators-svrj2\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:04 crc kubenswrapper[4990]: I1203 13:57:04.522863 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:05 crc kubenswrapper[4990]: I1203 13:57:05.108102 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svrj2"] Dec 03 13:57:05 crc kubenswrapper[4990]: I1203 13:57:05.880962 4990 generic.go:334] "Generic (PLEG): container finished" podID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerID="7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0" exitCode=0 Dec 03 13:57:05 crc kubenswrapper[4990]: I1203 13:57:05.881072 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerDied","Data":"7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0"} Dec 03 13:57:05 crc kubenswrapper[4990]: I1203 13:57:05.881526 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerStarted","Data":"c16e25f7681ec1f7fd25a69321b1b7b41a991e263cbc14d618ac48a18fd9e243"} Dec 03 13:57:06 crc kubenswrapper[4990]: I1203 13:57:06.893297 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerStarted","Data":"f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23"} Dec 03 13:57:07 crc kubenswrapper[4990]: I1203 13:57:07.904660 4990 generic.go:334] "Generic (PLEG): container finished" podID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerID="f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23" exitCode=0 Dec 03 13:57:07 crc kubenswrapper[4990]: I1203 13:57:07.904700 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerDied","Data":"f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23"} Dec 03 13:57:08 crc kubenswrapper[4990]: I1203 13:57:08.926509 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerStarted","Data":"a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e"} Dec 03 13:57:08 crc kubenswrapper[4990]: I1203 13:57:08.946874 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svrj2" podStartSLOduration=2.425635827 podStartE2EDuration="4.946856696s" podCreationTimestamp="2025-12-03 13:57:04 +0000 UTC" firstStartedPulling="2025-12-03 13:57:05.883372462 +0000 UTC m=+4774.025283691" lastFinishedPulling="2025-12-03 13:57:08.404593331 +0000 UTC m=+4776.546504560" observedRunningTime="2025-12-03 13:57:08.944683448 +0000 UTC m=+4777.086594677" watchObservedRunningTime="2025-12-03 13:57:08.946856696 +0000 UTC m=+4777.088767925" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.089244 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/util/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.329157 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/pull/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.398046 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/util/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.402318 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/pull/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.610768 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/util/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.623934 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/extract/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.640914 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmvrz8_9c828530-2366-4037-b371-69e9da5a0174/pull/0.log" Dec 03 13:57:09 crc kubenswrapper[4990]: I1203 13:57:09.810172 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/util/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.024793 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/util/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.031087 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/pull/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.057461 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/pull/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.288621 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/pull/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.304841 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/util/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.373396 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83jzsns_97bd525e-80e4-4b77-b2b6-e8d75c9ff66f/extract/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.531838 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-utilities/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.714179 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-utilities/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.767308 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-content/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.772911 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-content/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.873338 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-utilities/0.log" Dec 03 13:57:10 crc kubenswrapper[4990]: I1203 13:57:10.894041 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/extract-content/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.105734 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/extract-utilities/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.362621 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/extract-content/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.378582 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/extract-content/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.384281 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/extract-utilities/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.586860 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6xbrs_0b2e5cc6-9caa-458a-9f33-e0957c8fc435/registry-server/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.612767 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/registry-server/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.652327 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/extract-content/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.684598 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-svrj2_554c4066-d7ce-4bd5-be73-ebb33578a66c/extract-utilities/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.827410 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-utilities/0.log" Dec 03 13:57:11 crc kubenswrapper[4990]: I1203 13:57:11.979795 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-content/0.log" Dec 03 13:57:12 crc kubenswrapper[4990]: I1203 13:57:12.025666 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-utilities/0.log" Dec 03 13:57:12 crc kubenswrapper[4990]: I1203 13:57:12.035028 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-content/0.log" Dec 03 13:57:12 crc kubenswrapper[4990]: I1203 13:57:12.239502 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-content/0.log" Dec 03 13:57:12 crc kubenswrapper[4990]: I1203 13:57:12.286631 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/extract-utilities/0.log" Dec 03 13:57:12 crc kubenswrapper[4990]: I1203 13:57:12.502161 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9n6hl_8a0b82ca-6250-4ca2-abb5-2704907e4f98/marketplace-operator/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:12.618031 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-utilities/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.069833 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-content/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.134890 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zcrqv_88e0104c-98da-4fd8-86c6-5ba7a904d2b5/registry-server/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.154879 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-content/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.165709 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-utilities/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.311155 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-content/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.356117 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/extract-utilities/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.406159 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-utilities/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.517710 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pcfbg_2b52b9cb-d5fd-4916-a51b-440f95d47bbd/registry-server/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.586648 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-content/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.612989 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-utilities/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.630479 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-content/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.781597 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-content/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:13.784027 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/extract-utilities/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:14.395294 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h4nph_32457d7b-2669-40b2-8b4c-6d15db06c596/registry-server/0.log" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:14.523741 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:14.523793 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:14 crc kubenswrapper[4990]: I1203 13:57:14.574859 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:15 crc kubenswrapper[4990]: I1203 13:57:15.325253 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:15 crc kubenswrapper[4990]: I1203 13:57:15.377932 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svrj2"] Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.005841 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svrj2" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="registry-server" containerID="cri-o://a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e" gracePeriod=2 Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.454128 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.613139 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmk9d\" (UniqueName: \"kubernetes.io/projected/554c4066-d7ce-4bd5-be73-ebb33578a66c-kube-api-access-rmk9d\") pod \"554c4066-d7ce-4bd5-be73-ebb33578a66c\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.613250 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-catalog-content\") pod \"554c4066-d7ce-4bd5-be73-ebb33578a66c\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.613380 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-utilities\") pod \"554c4066-d7ce-4bd5-be73-ebb33578a66c\" (UID: \"554c4066-d7ce-4bd5-be73-ebb33578a66c\") " Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.614880 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-utilities" (OuterVolumeSpecName: "utilities") pod "554c4066-d7ce-4bd5-be73-ebb33578a66c" (UID: "554c4066-d7ce-4bd5-be73-ebb33578a66c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.616078 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.621691 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/554c4066-d7ce-4bd5-be73-ebb33578a66c-kube-api-access-rmk9d" (OuterVolumeSpecName: "kube-api-access-rmk9d") pod "554c4066-d7ce-4bd5-be73-ebb33578a66c" (UID: "554c4066-d7ce-4bd5-be73-ebb33578a66c"). InnerVolumeSpecName "kube-api-access-rmk9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.680046 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "554c4066-d7ce-4bd5-be73-ebb33578a66c" (UID: "554c4066-d7ce-4bd5-be73-ebb33578a66c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.717662 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmk9d\" (UniqueName: \"kubernetes.io/projected/554c4066-d7ce-4bd5-be73-ebb33578a66c-kube-api-access-rmk9d\") on node \"crc\" DevicePath \"\"" Dec 03 13:57:17 crc kubenswrapper[4990]: I1203 13:57:17.717719 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554c4066-d7ce-4bd5-be73-ebb33578a66c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.017886 4990 generic.go:334] "Generic (PLEG): container finished" podID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerID="a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e" exitCode=0 Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.017933 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerDied","Data":"a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e"} Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.017953 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svrj2" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.017974 4990 scope.go:117] "RemoveContainer" containerID="a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.017962 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svrj2" event={"ID":"554c4066-d7ce-4bd5-be73-ebb33578a66c","Type":"ContainerDied","Data":"c16e25f7681ec1f7fd25a69321b1b7b41a991e263cbc14d618ac48a18fd9e243"} Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.053527 4990 scope.go:117] "RemoveContainer" containerID="f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.059714 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svrj2"] Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.068631 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svrj2"] Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.083752 4990 scope.go:117] "RemoveContainer" containerID="7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.122366 4990 scope.go:117] "RemoveContainer" containerID="a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e" Dec 03 13:57:18 crc kubenswrapper[4990]: E1203 13:57:18.122895 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e\": container with ID starting with a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e not found: ID does not exist" containerID="a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.122937 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e"} err="failed to get container status \"a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e\": rpc error: code = NotFound desc = could not find container \"a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e\": container with ID starting with a00a759735fe16ef8de660ecb17392c6fe003a8dabf78660caec92442243721e not found: ID does not exist" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.122965 4990 scope.go:117] "RemoveContainer" containerID="f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23" Dec 03 13:57:18 crc kubenswrapper[4990]: E1203 13:57:18.124799 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23\": container with ID starting with f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23 not found: ID does not exist" containerID="f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.124927 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23"} err="failed to get container status \"f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23\": rpc error: code = NotFound desc = could not find container \"f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23\": container with ID starting with f72b64c7c9a022e0e03e3b8a7189ff75930899b2ad0e735e27b64ee17b1eff23 not found: ID does not exist" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.124942 4990 scope.go:117] "RemoveContainer" containerID="7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0" Dec 03 13:57:18 crc kubenswrapper[4990]: E1203 13:57:18.125395 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0\": container with ID starting with 7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0 not found: ID does not exist" containerID="7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.125434 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0"} err="failed to get container status \"7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0\": rpc error: code = NotFound desc = could not find container \"7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0\": container with ID starting with 7973c2e58696b9c46c03535af69b216b28350289ff057e0ca3cf295a78ffc5b0 not found: ID does not exist" Dec 03 13:57:18 crc kubenswrapper[4990]: I1203 13:57:18.275949 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" path="/var/lib/kubelet/pods/554c4066-d7ce-4bd5-be73-ebb33578a66c/volumes" Dec 03 13:58:33 crc kubenswrapper[4990]: I1203 13:58:33.286826 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:58:33 crc kubenswrapper[4990]: I1203 13:58:33.288524 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.254305 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxpdz"] Dec 03 13:58:51 crc kubenswrapper[4990]: E1203 13:58:51.256641 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="extract-content" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.256736 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="extract-content" Dec 03 13:58:51 crc kubenswrapper[4990]: E1203 13:58:51.256810 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="extract-utilities" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.256864 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="extract-utilities" Dec 03 13:58:51 crc kubenswrapper[4990]: E1203 13:58:51.256954 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="registry-server" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.257006 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="registry-server" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.257275 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="554c4066-d7ce-4bd5-be73-ebb33578a66c" containerName="registry-server" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.258865 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.274118 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlv9z\" (UniqueName: \"kubernetes.io/projected/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-kube-api-access-hlv9z\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.274262 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-catalog-content\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.274333 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-utilities\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.291267 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxpdz"] Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.375532 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlv9z\" (UniqueName: \"kubernetes.io/projected/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-kube-api-access-hlv9z\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.375616 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-catalog-content\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.375659 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-utilities\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.376228 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-catalog-content\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.376250 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-utilities\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.396302 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlv9z\" (UniqueName: \"kubernetes.io/projected/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-kube-api-access-hlv9z\") pod \"redhat-operators-pxpdz\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:51 crc kubenswrapper[4990]: I1203 13:58:51.582048 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:58:52 crc kubenswrapper[4990]: I1203 13:58:52.102014 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxpdz"] Dec 03 13:58:53 crc kubenswrapper[4990]: I1203 13:58:53.069403 4990 generic.go:334] "Generic (PLEG): container finished" podID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerID="20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb" exitCode=0 Dec 03 13:58:53 crc kubenswrapper[4990]: I1203 13:58:53.069514 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxpdz" event={"ID":"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c","Type":"ContainerDied","Data":"20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb"} Dec 03 13:58:53 crc kubenswrapper[4990]: I1203 13:58:53.069722 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxpdz" event={"ID":"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c","Type":"ContainerStarted","Data":"359869fceae59070aa5e8324bf574afbb9c055d51a36bb8d65550bc7ff5f11c4"} Dec 03 13:58:53 crc kubenswrapper[4990]: I1203 13:58:53.071656 4990 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 13:58:55 crc kubenswrapper[4990]: I1203 13:58:55.101174 4990 generic.go:334] "Generic (PLEG): container finished" podID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerID="62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30" exitCode=0 Dec 03 13:58:55 crc kubenswrapper[4990]: I1203 13:58:55.101295 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxpdz" event={"ID":"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c","Type":"ContainerDied","Data":"62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30"} Dec 03 13:58:57 crc kubenswrapper[4990]: I1203 13:58:57.121277 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxpdz" event={"ID":"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c","Type":"ContainerStarted","Data":"4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b"} Dec 03 13:58:57 crc kubenswrapper[4990]: I1203 13:58:57.149761 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxpdz" podStartSLOduration=3.4730081090000002 podStartE2EDuration="6.149735713s" podCreationTimestamp="2025-12-03 13:58:51 +0000 UTC" firstStartedPulling="2025-12-03 13:58:53.071369305 +0000 UTC m=+4881.213280534" lastFinishedPulling="2025-12-03 13:58:55.748096909 +0000 UTC m=+4883.890008138" observedRunningTime="2025-12-03 13:58:57.140727263 +0000 UTC m=+4885.282638502" watchObservedRunningTime="2025-12-03 13:58:57.149735713 +0000 UTC m=+4885.291646942" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.080897 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-76thw"] Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.083373 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.097620 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-76thw"] Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.176704 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-utilities\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.176773 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpg9h\" (UniqueName: \"kubernetes.io/projected/f3d25817-1992-424e-8d2f-36fcb0adf845-kube-api-access-gpg9h\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.177267 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-catalog-content\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.278778 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-catalog-content\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.278907 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-utilities\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.278940 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpg9h\" (UniqueName: \"kubernetes.io/projected/f3d25817-1992-424e-8d2f-36fcb0adf845-kube-api-access-gpg9h\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.279425 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-utilities\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.280337 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-catalog-content\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.301642 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpg9h\" (UniqueName: \"kubernetes.io/projected/f3d25817-1992-424e-8d2f-36fcb0adf845-kube-api-access-gpg9h\") pod \"certified-operators-76thw\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.413027 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.583205 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.583536 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.658952 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:59:01 crc kubenswrapper[4990]: I1203 13:59:01.955313 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-76thw"] Dec 03 13:59:02 crc kubenswrapper[4990]: I1203 13:59:02.174008 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerStarted","Data":"fa8356533e9a84244341ef2dd1c2bbaece6a21daad89de9ae042e075e6acba05"} Dec 03 13:59:02 crc kubenswrapper[4990]: I1203 13:59:02.220851 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:59:03 crc kubenswrapper[4990]: I1203 13:59:03.287176 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:59:03 crc kubenswrapper[4990]: I1203 13:59:03.287270 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.053715 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxpdz"] Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.194350 4990 generic.go:334] "Generic (PLEG): container finished" podID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerID="47b87c38d1bce7629ddab6cac28acfb8c7278c0a60fdb840691374f1eba379b8" exitCode=0 Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.194435 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx545/must-gather-vkd6g" event={"ID":"63813ce7-5ad3-45b3-a488-9de161b199c3","Type":"ContainerDied","Data":"47b87c38d1bce7629ddab6cac28acfb8c7278c0a60fdb840691374f1eba379b8"} Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.195388 4990 scope.go:117] "RemoveContainer" containerID="47b87c38d1bce7629ddab6cac28acfb8c7278c0a60fdb840691374f1eba379b8" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.196237 4990 generic.go:334] "Generic (PLEG): container finished" podID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerID="c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f" exitCode=0 Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.196335 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerDied","Data":"c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f"} Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.196619 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pxpdz" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="registry-server" containerID="cri-o://4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b" gracePeriod=2 Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.600339 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.758521 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-utilities\") pod \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.758741 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlv9z\" (UniqueName: \"kubernetes.io/projected/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-kube-api-access-hlv9z\") pod \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.758889 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-catalog-content\") pod \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\" (UID: \"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c\") " Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.759425 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-utilities" (OuterVolumeSpecName: "utilities") pod "900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" (UID: "900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.763949 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-kube-api-access-hlv9z" (OuterVolumeSpecName: "kube-api-access-hlv9z") pod "900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" (UID: "900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c"). InnerVolumeSpecName "kube-api-access-hlv9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.862688 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.862740 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlv9z\" (UniqueName: \"kubernetes.io/projected/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-kube-api-access-hlv9z\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.887546 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" (UID: "900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.964878 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:04 crc kubenswrapper[4990]: I1203 13:59:04.976071 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx545_must-gather-vkd6g_63813ce7-5ad3-45b3-a488-9de161b199c3/gather/0.log" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.208467 4990 generic.go:334] "Generic (PLEG): container finished" podID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerID="4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b" exitCode=0 Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.208536 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxpdz" event={"ID":"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c","Type":"ContainerDied","Data":"4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b"} Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.208593 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxpdz" event={"ID":"900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c","Type":"ContainerDied","Data":"359869fceae59070aa5e8324bf574afbb9c055d51a36bb8d65550bc7ff5f11c4"} Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.208619 4990 scope.go:117] "RemoveContainer" containerID="4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.208559 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxpdz" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.212720 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerStarted","Data":"086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50"} Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.232382 4990 scope.go:117] "RemoveContainer" containerID="62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.268172 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxpdz"] Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.269157 4990 scope.go:117] "RemoveContainer" containerID="20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.279308 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pxpdz"] Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.290477 4990 scope.go:117] "RemoveContainer" containerID="4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b" Dec 03 13:59:05 crc kubenswrapper[4990]: E1203 13:59:05.291134 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b\": container with ID starting with 4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b not found: ID does not exist" containerID="4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.291191 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b"} err="failed to get container status \"4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b\": rpc error: code = NotFound desc = could not find container \"4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b\": container with ID starting with 4a08bc32f700f1d2d8462dddd36ed31aae5e2c700374b0af2ff0935d3ab7f80b not found: ID does not exist" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.291209 4990 scope.go:117] "RemoveContainer" containerID="62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30" Dec 03 13:59:05 crc kubenswrapper[4990]: E1203 13:59:05.292100 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30\": container with ID starting with 62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30 not found: ID does not exist" containerID="62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.292152 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30"} err="failed to get container status \"62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30\": rpc error: code = NotFound desc = could not find container \"62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30\": container with ID starting with 62cc2082afc449aab33628bf81943d9c65164e7e48e2a79fc70185649ebc4f30 not found: ID does not exist" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.292182 4990 scope.go:117] "RemoveContainer" containerID="20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb" Dec 03 13:59:05 crc kubenswrapper[4990]: E1203 13:59:05.292658 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb\": container with ID starting with 20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb not found: ID does not exist" containerID="20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb" Dec 03 13:59:05 crc kubenswrapper[4990]: I1203 13:59:05.292690 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb"} err="failed to get container status \"20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb\": rpc error: code = NotFound desc = could not find container \"20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb\": container with ID starting with 20d7892e85cba316feaf210847a957019c06a1724ff59d1dea6d4dfb52e3d6fb not found: ID does not exist" Dec 03 13:59:06 crc kubenswrapper[4990]: I1203 13:59:06.225312 4990 generic.go:334] "Generic (PLEG): container finished" podID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerID="086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50" exitCode=0 Dec 03 13:59:06 crc kubenswrapper[4990]: I1203 13:59:06.225392 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerDied","Data":"086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50"} Dec 03 13:59:06 crc kubenswrapper[4990]: I1203 13:59:06.278536 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" path="/var/lib/kubelet/pods/900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c/volumes" Dec 03 13:59:08 crc kubenswrapper[4990]: I1203 13:59:08.247183 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerStarted","Data":"0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313"} Dec 03 13:59:08 crc kubenswrapper[4990]: I1203 13:59:08.283618 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-76thw" podStartSLOduration=4.897082964 podStartE2EDuration="7.283585231s" podCreationTimestamp="2025-12-03 13:59:01 +0000 UTC" firstStartedPulling="2025-12-03 13:59:04.197852367 +0000 UTC m=+4892.339763596" lastFinishedPulling="2025-12-03 13:59:06.584354634 +0000 UTC m=+4894.726265863" observedRunningTime="2025-12-03 13:59:08.267385359 +0000 UTC m=+4896.409296578" watchObservedRunningTime="2025-12-03 13:59:08.283585231 +0000 UTC m=+4896.425496480" Dec 03 13:59:11 crc kubenswrapper[4990]: I1203 13:59:11.413790 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:11 crc kubenswrapper[4990]: I1203 13:59:11.414124 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:11 crc kubenswrapper[4990]: I1203 13:59:11.463646 4990 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:12 crc kubenswrapper[4990]: I1203 13:59:12.339039 4990 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:12 crc kubenswrapper[4990]: I1203 13:59:12.386217 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-76thw"] Dec 03 13:59:14 crc kubenswrapper[4990]: I1203 13:59:14.313910 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-76thw" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="registry-server" containerID="cri-o://0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313" gracePeriod=2 Dec 03 13:59:14 crc kubenswrapper[4990]: I1203 13:59:14.408048 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx545/must-gather-vkd6g"] Dec 03 13:59:14 crc kubenswrapper[4990]: I1203 13:59:14.408367 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zx545/must-gather-vkd6g" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="copy" containerID="cri-o://a0aa0f9019012eef297c823e52f6359c00f6f72e325e6bab90282fdd2b85b531" gracePeriod=2 Dec 03 13:59:14 crc kubenswrapper[4990]: I1203 13:59:14.419485 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx545/must-gather-vkd6g"] Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.330787 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.337147 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx545_must-gather-vkd6g_63813ce7-5ad3-45b3-a488-9de161b199c3/copy/0.log" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.339371 4990 generic.go:334] "Generic (PLEG): container finished" podID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerID="a0aa0f9019012eef297c823e52f6359c00f6f72e325e6bab90282fdd2b85b531" exitCode=143 Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.344296 4990 generic.go:334] "Generic (PLEG): container finished" podID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerID="0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313" exitCode=0 Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.344348 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerDied","Data":"0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313"} Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.344405 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76thw" event={"ID":"f3d25817-1992-424e-8d2f-36fcb0adf845","Type":"ContainerDied","Data":"fa8356533e9a84244341ef2dd1c2bbaece6a21daad89de9ae042e075e6acba05"} Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.344470 4990 scope.go:117] "RemoveContainer" containerID="0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.344718 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76thw" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.370409 4990 scope.go:117] "RemoveContainer" containerID="086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.438627 4990 scope.go:117] "RemoveContainer" containerID="c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.461792 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-catalog-content\") pod \"f3d25817-1992-424e-8d2f-36fcb0adf845\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.462007 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-utilities\") pod \"f3d25817-1992-424e-8d2f-36fcb0adf845\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.462093 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpg9h\" (UniqueName: \"kubernetes.io/projected/f3d25817-1992-424e-8d2f-36fcb0adf845-kube-api-access-gpg9h\") pod \"f3d25817-1992-424e-8d2f-36fcb0adf845\" (UID: \"f3d25817-1992-424e-8d2f-36fcb0adf845\") " Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.463866 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-utilities" (OuterVolumeSpecName: "utilities") pod "f3d25817-1992-424e-8d2f-36fcb0adf845" (UID: "f3d25817-1992-424e-8d2f-36fcb0adf845"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.469563 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d25817-1992-424e-8d2f-36fcb0adf845-kube-api-access-gpg9h" (OuterVolumeSpecName: "kube-api-access-gpg9h") pod "f3d25817-1992-424e-8d2f-36fcb0adf845" (UID: "f3d25817-1992-424e-8d2f-36fcb0adf845"). InnerVolumeSpecName "kube-api-access-gpg9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.531976 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3d25817-1992-424e-8d2f-36fcb0adf845" (UID: "f3d25817-1992-424e-8d2f-36fcb0adf845"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.544060 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx545_must-gather-vkd6g_63813ce7-5ad3-45b3-a488-9de161b199c3/copy/0.log" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.544426 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.555726 4990 scope.go:117] "RemoveContainer" containerID="0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313" Dec 03 13:59:15 crc kubenswrapper[4990]: E1203 13:59:15.556708 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313\": container with ID starting with 0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313 not found: ID does not exist" containerID="0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.556765 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313"} err="failed to get container status \"0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313\": rpc error: code = NotFound desc = could not find container \"0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313\": container with ID starting with 0138082e708f2ff3d4311636c5fb53b6de0e09a66100637e54823bd734f99313 not found: ID does not exist" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.556816 4990 scope.go:117] "RemoveContainer" containerID="086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50" Dec 03 13:59:15 crc kubenswrapper[4990]: E1203 13:59:15.557139 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50\": container with ID starting with 086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50 not found: ID does not exist" containerID="086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.557170 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50"} err="failed to get container status \"086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50\": rpc error: code = NotFound desc = could not find container \"086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50\": container with ID starting with 086a264f917896db4d268c6bd6062f4a09bf2d24717d7d1cf522824cd8797c50 not found: ID does not exist" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.557189 4990 scope.go:117] "RemoveContainer" containerID="c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f" Dec 03 13:59:15 crc kubenswrapper[4990]: E1203 13:59:15.557672 4990 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f\": container with ID starting with c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f not found: ID does not exist" containerID="c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.557704 4990 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f"} err="failed to get container status \"c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f\": rpc error: code = NotFound desc = could not find container \"c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f\": container with ID starting with c3f0e25ecc996f2534fb30d99d62877f8846e048851d06288c412b6130f9368f not found: ID does not exist" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.565312 4990 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.565355 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpg9h\" (UniqueName: \"kubernetes.io/projected/f3d25817-1992-424e-8d2f-36fcb0adf845-kube-api-access-gpg9h\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.565370 4990 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d25817-1992-424e-8d2f-36fcb0adf845-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.666250 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/63813ce7-5ad3-45b3-a488-9de161b199c3-must-gather-output\") pod \"63813ce7-5ad3-45b3-a488-9de161b199c3\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.666423 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-779q5\" (UniqueName: \"kubernetes.io/projected/63813ce7-5ad3-45b3-a488-9de161b199c3-kube-api-access-779q5\") pod \"63813ce7-5ad3-45b3-a488-9de161b199c3\" (UID: \"63813ce7-5ad3-45b3-a488-9de161b199c3\") " Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.671964 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63813ce7-5ad3-45b3-a488-9de161b199c3-kube-api-access-779q5" (OuterVolumeSpecName: "kube-api-access-779q5") pod "63813ce7-5ad3-45b3-a488-9de161b199c3" (UID: "63813ce7-5ad3-45b3-a488-9de161b199c3"). InnerVolumeSpecName "kube-api-access-779q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.688313 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-76thw"] Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.695876 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-76thw"] Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.768976 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-779q5\" (UniqueName: \"kubernetes.io/projected/63813ce7-5ad3-45b3-a488-9de161b199c3-kube-api-access-779q5\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.818371 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63813ce7-5ad3-45b3-a488-9de161b199c3-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "63813ce7-5ad3-45b3-a488-9de161b199c3" (UID: "63813ce7-5ad3-45b3-a488-9de161b199c3"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 13:59:15 crc kubenswrapper[4990]: I1203 13:59:15.870857 4990 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/63813ce7-5ad3-45b3-a488-9de161b199c3-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 13:59:16 crc kubenswrapper[4990]: I1203 13:59:16.276648 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" path="/var/lib/kubelet/pods/63813ce7-5ad3-45b3-a488-9de161b199c3/volumes" Dec 03 13:59:16 crc kubenswrapper[4990]: I1203 13:59:16.277557 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" path="/var/lib/kubelet/pods/f3d25817-1992-424e-8d2f-36fcb0adf845/volumes" Dec 03 13:59:16 crc kubenswrapper[4990]: I1203 13:59:16.356898 4990 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx545_must-gather-vkd6g_63813ce7-5ad3-45b3-a488-9de161b199c3/copy/0.log" Dec 03 13:59:16 crc kubenswrapper[4990]: I1203 13:59:16.357766 4990 scope.go:117] "RemoveContainer" containerID="a0aa0f9019012eef297c823e52f6359c00f6f72e325e6bab90282fdd2b85b531" Dec 03 13:59:16 crc kubenswrapper[4990]: I1203 13:59:16.357789 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx545/must-gather-vkd6g" Dec 03 13:59:16 crc kubenswrapper[4990]: I1203 13:59:16.378140 4990 scope.go:117] "RemoveContainer" containerID="47b87c38d1bce7629ddab6cac28acfb8c7278c0a60fdb840691374f1eba379b8" Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.286690 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.288465 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.288661 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.289951 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28c51c6562bf6ec0845cefe19910d3a93dce4c1ca0c2fd94d19a2e02f1f60412"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.290089 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://28c51c6562bf6ec0845cefe19910d3a93dce4c1ca0c2fd94d19a2e02f1f60412" gracePeriod=600 Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.522492 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="28c51c6562bf6ec0845cefe19910d3a93dce4c1ca0c2fd94d19a2e02f1f60412" exitCode=0 Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.522556 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"28c51c6562bf6ec0845cefe19910d3a93dce4c1ca0c2fd94d19a2e02f1f60412"} Dec 03 13:59:33 crc kubenswrapper[4990]: I1203 13:59:33.522659 4990 scope.go:117] "RemoveContainer" containerID="4b375a6898824a288ed4082080cdcf13f00aa7032c843f320b52a2e0eb25093f" Dec 03 13:59:34 crc kubenswrapper[4990]: I1203 13:59:34.534176 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerStarted","Data":"7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46"} Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.145795 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct"] Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146777 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="gather" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146793 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="gather" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146803 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="extract-utilities" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146810 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="extract-utilities" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146823 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="extract-content" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146831 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="extract-content" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146846 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="extract-utilities" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146851 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="extract-utilities" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146874 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="copy" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146880 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="copy" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146888 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146895 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146910 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146915 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4990]: E1203 14:00:00.146929 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="extract-content" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.146934 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="extract-content" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.147160 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="copy" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.147172 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="900a4b4f-d94b-40f2-975e-4d9d3bd6ba7c" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.147182 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="63813ce7-5ad3-45b3-a488-9de161b199c3" containerName="gather" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.147202 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d25817-1992-424e-8d2f-36fcb0adf845" containerName="registry-server" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.147907 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.149589 4990 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.149843 4990 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.155526 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct"] Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.250849 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-config-volume\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.251233 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-secret-volume\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.251382 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b7j4\" (UniqueName: \"kubernetes.io/projected/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-kube-api-access-8b7j4\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.353240 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-secret-volume\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.353374 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b7j4\" (UniqueName: \"kubernetes.io/projected/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-kube-api-access-8b7j4\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.353646 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-config-volume\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.355013 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-config-volume\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.363165 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-secret-volume\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.369906 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b7j4\" (UniqueName: \"kubernetes.io/projected/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-kube-api-access-8b7j4\") pod \"collect-profiles-29412840-x4wct\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.470731 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:00 crc kubenswrapper[4990]: W1203 14:00:00.925684 4990 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc68d4a1e_7a2a_48c9_9920_c1d6b37f759e.slice/crio-1a386892d58baf40d9655fc174245a7523afdf86f6bbc924bf9670682fd5f4f5 WatchSource:0}: Error finding container 1a386892d58baf40d9655fc174245a7523afdf86f6bbc924bf9670682fd5f4f5: Status 404 returned error can't find the container with id 1a386892d58baf40d9655fc174245a7523afdf86f6bbc924bf9670682fd5f4f5 Dec 03 14:00:00 crc kubenswrapper[4990]: I1203 14:00:00.925904 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct"] Dec 03 14:00:01 crc kubenswrapper[4990]: I1203 14:00:01.760254 4990 generic.go:334] "Generic (PLEG): container finished" podID="c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" containerID="ea571ce5db8f4008332fa0a8a2d2d62696be4baf590363a9a4491e3e9303618e" exitCode=0 Dec 03 14:00:01 crc kubenswrapper[4990]: I1203 14:00:01.760316 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" event={"ID":"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e","Type":"ContainerDied","Data":"ea571ce5db8f4008332fa0a8a2d2d62696be4baf590363a9a4491e3e9303618e"} Dec 03 14:00:01 crc kubenswrapper[4990]: I1203 14:00:01.760760 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" event={"ID":"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e","Type":"ContainerStarted","Data":"1a386892d58baf40d9655fc174245a7523afdf86f6bbc924bf9670682fd5f4f5"} Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.118131 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.210107 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-secret-volume\") pod \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.210193 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b7j4\" (UniqueName: \"kubernetes.io/projected/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-kube-api-access-8b7j4\") pod \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.210504 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-config-volume\") pod \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\" (UID: \"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e\") " Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.211253 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" (UID: "c68d4a1e-7a2a-48c9-9920-c1d6b37f759e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.216482 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-kube-api-access-8b7j4" (OuterVolumeSpecName: "kube-api-access-8b7j4") pod "c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" (UID: "c68d4a1e-7a2a-48c9-9920-c1d6b37f759e"). InnerVolumeSpecName "kube-api-access-8b7j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.217142 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" (UID: "c68d4a1e-7a2a-48c9-9920-c1d6b37f759e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.313485 4990 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.313540 4990 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.313560 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b7j4\" (UniqueName: \"kubernetes.io/projected/c68d4a1e-7a2a-48c9-9920-c1d6b37f759e-kube-api-access-8b7j4\") on node \"crc\" DevicePath \"\"" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.782543 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" event={"ID":"c68d4a1e-7a2a-48c9-9920-c1d6b37f759e","Type":"ContainerDied","Data":"1a386892d58baf40d9655fc174245a7523afdf86f6bbc924bf9670682fd5f4f5"} Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.782594 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a386892d58baf40d9655fc174245a7523afdf86f6bbc924bf9670682fd5f4f5" Dec 03 14:00:03 crc kubenswrapper[4990]: I1203 14:00:03.782602 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412840-x4wct" Dec 03 14:00:04 crc kubenswrapper[4990]: I1203 14:00:04.218220 4990 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4"] Dec 03 14:00:04 crc kubenswrapper[4990]: I1203 14:00:04.228177 4990 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412795-8ggb4"] Dec 03 14:00:04 crc kubenswrapper[4990]: I1203 14:00:04.275100 4990 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a53162-4ed1-48f6-bdf2-4dc08f7ba143" path="/var/lib/kubelet/pods/22a53162-4ed1-48f6-bdf2-4dc08f7ba143/volumes" Dec 03 14:00:27 crc kubenswrapper[4990]: I1203 14:00:27.677125 4990 scope.go:117] "RemoveContainer" containerID="99dde99af72854895d0943426970d10b72e0fc1c528a197a7b0bb13f4328134c" Dec 03 14:00:27 crc kubenswrapper[4990]: I1203 14:00:27.709243 4990 scope.go:117] "RemoveContainer" containerID="6baa2e1e5f3d7af054d9629c36642024cff71374a31dbb911267a5d7be325814" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.151416 4990 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412841-pvp76"] Dec 03 14:01:00 crc kubenswrapper[4990]: E1203 14:01:00.154084 4990 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" containerName="collect-profiles" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.154107 4990 state_mem.go:107] "Deleted CPUSet assignment" podUID="c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" containerName="collect-profiles" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.154386 4990 memory_manager.go:354] "RemoveStaleState removing state" podUID="c68d4a1e-7a2a-48c9-9920-c1d6b37f759e" containerName="collect-profiles" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.155317 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.160938 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412841-pvp76"] Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.251428 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-config-data\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.251845 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-fernet-keys\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.252204 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twff4\" (UniqueName: \"kubernetes.io/projected/bcccee8f-b506-4f89-8a8a-9b6857be55f3-kube-api-access-twff4\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.252579 4990 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-combined-ca-bundle\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.354744 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-fernet-keys\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.354847 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twff4\" (UniqueName: \"kubernetes.io/projected/bcccee8f-b506-4f89-8a8a-9b6857be55f3-kube-api-access-twff4\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.354933 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-combined-ca-bundle\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.354970 4990 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-config-data\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.361035 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-combined-ca-bundle\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.361613 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-fernet-keys\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.363369 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-config-data\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.369903 4990 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twff4\" (UniqueName: \"kubernetes.io/projected/bcccee8f-b506-4f89-8a8a-9b6857be55f3-kube-api-access-twff4\") pod \"keystone-cron-29412841-pvp76\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.482831 4990 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:00 crc kubenswrapper[4990]: I1203 14:01:00.896413 4990 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412841-pvp76"] Dec 03 14:01:01 crc kubenswrapper[4990]: I1203 14:01:01.328986 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-pvp76" event={"ID":"bcccee8f-b506-4f89-8a8a-9b6857be55f3","Type":"ContainerStarted","Data":"e052f7b371693e29326b296a2ed43e3ddfa2228236651c3f4462e319edf54048"} Dec 03 14:01:01 crc kubenswrapper[4990]: I1203 14:01:01.329426 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-pvp76" event={"ID":"bcccee8f-b506-4f89-8a8a-9b6857be55f3","Type":"ContainerStarted","Data":"de253015c73867cd7fc152f694fc79ae8d23fd17eb326af186cfb3b29998e6ae"} Dec 03 14:01:01 crc kubenswrapper[4990]: I1203 14:01:01.353756 4990 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412841-pvp76" podStartSLOduration=1.353734341 podStartE2EDuration="1.353734341s" podCreationTimestamp="2025-12-03 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 14:01:01.347388694 +0000 UTC m=+5009.489299933" watchObservedRunningTime="2025-12-03 14:01:01.353734341 +0000 UTC m=+5009.495645570" Dec 03 14:01:07 crc kubenswrapper[4990]: I1203 14:01:07.398250 4990 generic.go:334] "Generic (PLEG): container finished" podID="bcccee8f-b506-4f89-8a8a-9b6857be55f3" containerID="e052f7b371693e29326b296a2ed43e3ddfa2228236651c3f4462e319edf54048" exitCode=0 Dec 03 14:01:07 crc kubenswrapper[4990]: I1203 14:01:07.398336 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-pvp76" event={"ID":"bcccee8f-b506-4f89-8a8a-9b6857be55f3","Type":"ContainerDied","Data":"e052f7b371693e29326b296a2ed43e3ddfa2228236651c3f4462e319edf54048"} Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.765399 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.819297 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-combined-ca-bundle\") pod \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.819549 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-config-data\") pod \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.819578 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-fernet-keys\") pod \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.819650 4990 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twff4\" (UniqueName: \"kubernetes.io/projected/bcccee8f-b506-4f89-8a8a-9b6857be55f3-kube-api-access-twff4\") pod \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\" (UID: \"bcccee8f-b506-4f89-8a8a-9b6857be55f3\") " Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.825940 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcccee8f-b506-4f89-8a8a-9b6857be55f3-kube-api-access-twff4" (OuterVolumeSpecName: "kube-api-access-twff4") pod "bcccee8f-b506-4f89-8a8a-9b6857be55f3" (UID: "bcccee8f-b506-4f89-8a8a-9b6857be55f3"). InnerVolumeSpecName "kube-api-access-twff4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.826050 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bcccee8f-b506-4f89-8a8a-9b6857be55f3" (UID: "bcccee8f-b506-4f89-8a8a-9b6857be55f3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.858158 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcccee8f-b506-4f89-8a8a-9b6857be55f3" (UID: "bcccee8f-b506-4f89-8a8a-9b6857be55f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.883792 4990 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-config-data" (OuterVolumeSpecName: "config-data") pod "bcccee8f-b506-4f89-8a8a-9b6857be55f3" (UID: "bcccee8f-b506-4f89-8a8a-9b6857be55f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.922060 4990 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.922091 4990 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.922102 4990 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twff4\" (UniqueName: \"kubernetes.io/projected/bcccee8f-b506-4f89-8a8a-9b6857be55f3-kube-api-access-twff4\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:08 crc kubenswrapper[4990]: I1203 14:01:08.922124 4990 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcccee8f-b506-4f89-8a8a-9b6857be55f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 14:01:09 crc kubenswrapper[4990]: I1203 14:01:09.416823 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412841-pvp76" event={"ID":"bcccee8f-b506-4f89-8a8a-9b6857be55f3","Type":"ContainerDied","Data":"de253015c73867cd7fc152f694fc79ae8d23fd17eb326af186cfb3b29998e6ae"} Dec 03 14:01:09 crc kubenswrapper[4990]: I1203 14:01:09.416874 4990 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de253015c73867cd7fc152f694fc79ae8d23fd17eb326af186cfb3b29998e6ae" Dec 03 14:01:09 crc kubenswrapper[4990]: I1203 14:01:09.416876 4990 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412841-pvp76" Dec 03 14:01:33 crc kubenswrapper[4990]: I1203 14:01:33.286562 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:01:33 crc kubenswrapper[4990]: I1203 14:01:33.287162 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:02:03 crc kubenswrapper[4990]: I1203 14:02:03.286840 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:02:03 crc kubenswrapper[4990]: I1203 14:02:03.287399 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:02:33 crc kubenswrapper[4990]: I1203 14:02:33.286657 4990 patch_prober.go:28] interesting pod/machine-config-daemon-85qrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 14:02:33 crc kubenswrapper[4990]: I1203 14:02:33.287095 4990 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 14:02:33 crc kubenswrapper[4990]: I1203 14:02:33.287140 4990 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" Dec 03 14:02:33 crc kubenswrapper[4990]: I1203 14:02:33.287896 4990 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46"} pod="openshift-machine-config-operator/machine-config-daemon-85qrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 14:02:33 crc kubenswrapper[4990]: I1203 14:02:33.287947 4990 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerName="machine-config-daemon" containerID="cri-o://7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46" gracePeriod=600 Dec 03 14:02:33 crc kubenswrapper[4990]: E1203 14:02:33.425343 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 14:02:34 crc kubenswrapper[4990]: I1203 14:02:34.440236 4990 generic.go:334] "Generic (PLEG): container finished" podID="7bec8561-741c-4adf-bfce-6f80a0dfb170" containerID="7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46" exitCode=0 Dec 03 14:02:34 crc kubenswrapper[4990]: I1203 14:02:34.440432 4990 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" event={"ID":"7bec8561-741c-4adf-bfce-6f80a0dfb170","Type":"ContainerDied","Data":"7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46"} Dec 03 14:02:34 crc kubenswrapper[4990]: I1203 14:02:34.440590 4990 scope.go:117] "RemoveContainer" containerID="28c51c6562bf6ec0845cefe19910d3a93dce4c1ca0c2fd94d19a2e02f1f60412" Dec 03 14:02:34 crc kubenswrapper[4990]: I1203 14:02:34.441334 4990 scope.go:117] "RemoveContainer" containerID="7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46" Dec 03 14:02:34 crc kubenswrapper[4990]: E1203 14:02:34.441693 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 14:02:47 crc kubenswrapper[4990]: I1203 14:02:47.265146 4990 scope.go:117] "RemoveContainer" containerID="7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46" Dec 03 14:02:47 crc kubenswrapper[4990]: E1203 14:02:47.265887 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 14:02:58 crc kubenswrapper[4990]: I1203 14:02:58.265004 4990 scope.go:117] "RemoveContainer" containerID="7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46" Dec 03 14:02:58 crc kubenswrapper[4990]: E1203 14:02:58.265816 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" Dec 03 14:03:12 crc kubenswrapper[4990]: I1203 14:03:12.271356 4990 scope.go:117] "RemoveContainer" containerID="7c38e74db88ed57a8b253e6159afe56174bae976c96553ab020403985574ab46" Dec 03 14:03:12 crc kubenswrapper[4990]: E1203 14:03:12.272118 4990 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-85qrd_openshift-machine-config-operator(7bec8561-741c-4adf-bfce-6f80a0dfb170)\"" pod="openshift-machine-config-operator/machine-config-daemon-85qrd" podUID="7bec8561-741c-4adf-bfce-6f80a0dfb170" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114041457024447 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114041460017356 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114027053016503 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114027053015453 5ustar corecore